• 0 Posts
  • 17 Comments
Joined 4 months ago
cake
Cake day: March 3rd, 2024

help-circle


  • Understanding the variety of speech over a drive-thru speaker can be difficult for a human with experience in the job. I can’t see the current level of voice recognition matching it, especially if it’s using LLMs for processing of what it managed to detect. If I’m placing a food order I don’t need a LLM hallucination to try and fill in blanks of what it didn’t convert correctly to tokens or wasn’t trained on.





  • Is it a physical HD (magnetic) and making noise? I had one years ago (fortunately my only failure so far) and if I kept persisting to try and read it via a USB recovery drive, I managed to pull enough data off that was important. If it’s a newer SSD, that’s a different thing. Doesn’t mean all the data is gone, just a lot harder (read $$$) to pull. Hopefully it’s just software or a loose cable.




  • The narrow purpose models seem to be the most successful, so this would support the idea that a general AI isn’t going to happen from LLMs alone. It’s interesting that hallucinations are seen as a problem yet are probably part of why LLMs can be creative (much like humans). We shouldn’t want to stop them, but just control when they happen and be aware of when the AI is off the tracks. A group of different models working together and checking each other might work (and probably has already been tried, it’s hard to keep up).