• 0 Posts
  • 65 Comments
Joined 4 months ago
cake
Cake day: March 3rd, 2024

help-circle




  • I’ve come to the conclusion that all these breach notices and the free stuff they offer for X months is a huge scam to get you sign up up for something. Either that, or every company has woefully underpaid/incompetent IT people. I’m waiting for the next news story to break on another company that somehow got passwords or identity info hacked that was stored in plain text…something I learned how to not do back in the 90s with basic HTML and PHP.

    In short - I don’t believe them. They all are using the same form letters, it’s a scheme that they’re all in on.








  • Understanding the variety of speech over a drive-thru speaker can be difficult for a human with experience in the job. I can’t see the current level of voice recognition matching it, especially if it’s using LLMs for processing of what it managed to detect. If I’m placing a food order I don’t need a LLM hallucination to try and fill in blanks of what it didn’t convert correctly to tokens or wasn’t trained on.





  • Is it a physical HD (magnetic) and making noise? I had one years ago (fortunately my only failure so far) and if I kept persisting to try and read it via a USB recovery drive, I managed to pull enough data off that was important. If it’s a newer SSD, that’s a different thing. Doesn’t mean all the data is gone, just a lot harder (read $$$) to pull. Hopefully it’s just software or a loose cable.




  • The narrow purpose models seem to be the most successful, so this would support the idea that a general AI isn’t going to happen from LLMs alone. It’s interesting that hallucinations are seen as a problem yet are probably part of why LLMs can be creative (much like humans). We shouldn’t want to stop them, but just control when they happen and be aware of when the AI is off the tracks. A group of different models working together and checking each other might work (and probably has already been tried, it’s hard to keep up).