this post was submitted on 29 Apr 2024
195 points (94.9% liked)
Technology
59157 readers
2348 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
It "knows" as in it has access to the information and the ability to provide the right info for the right context.
Any part of that process the AI can just "bullshit" and fills in the gaps with random stuff.
Which is what you want when it's "learning". You want it to try so it's attempt can be rated, and the relevant info added to its "knowledge".
But when consumers are using it, you want it to say "I can't answer that". But consumers are usually stupid and will buy/use the one that says "I can't answer that" the least.
Which is why AI should tell end users "I don't know" more often.
If you feel this is a simple solution, I strongly suggest you write up exactly how you do this and make yourself a billion dollars.
It doesn't, though, any more than you have access to the information in a pile of 10 million shredded documents.
Right, in this case that we're talking about...
Do you not understand how "answer unavailable" is a better answer than taking a small percent of strips of paper at random and filling in the rest with words that sound relevant?
It's like a mad libs
Right. They're text generators. That's the technology. It can't do what you're demanding because that's not how it works. LLMs aren't magic answer machines. They don't know when to say "answer not available". They don't know what they're being asked. They don't know anything.
That is what LLMs do in EVERY conversation. Most of the time you don't notice it, because it fits your expectations.
You know that answer unavailable is better because you have real intelligence, an LLM is just some mathematical functions so it can't do that. If it could it would be getting much closer to actually being AI.