this post was submitted on 23 Mar 2025
772 points (97.8% liked)
Technology
67669 readers
5714 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
It's AI. There's nothing to delete but the erroneous response. There is no database of facts to edit. It doesn't know fact from fiction, and the response is also very much skewed by the context of the query. I could easily get it to say the same about nearly any random name just by asking it about a bunch of family murders and then asking about a name it doesn't recognize. It is more likely to assume that person is in the same category as the others and if the one or more of the names have any association (real or fictional) with murder.
The fact you chose to make your data storage unreadable, doesn't relieve you of the responsibilities inherent to storing the data.
Throwing away my car key won't protect me from paying parking tickets i accrue while being physically unable to move my car.
It's not unreadable, it doesn't exist.
The responses are just statistically what sounds vaugly what you want to hear.
They can erase the chat responses, but that won't stop it from generating it again.
Generative AI doesn't start with facts and work from there. It's just statistically what you want to hear.
Then what do you mean trained AI models are?
The ai model is trained on data and encodes unknown parts of that data in its weights.
This is data storage. Unmanageable, almost unknowable data storage, but still data storage.
If it didn't store data it couldn't learn from its training.
Your still placing more intent and facts into those processes than actually exist.
You cant even get it to count how many letter p are in the word apple. At least not last time I tried.
That storage your talking about isn't facts. It's how sentences are structured and what they "mean".
As for the output "meaning" it's still just guessing what you want to hear. No facts involved.
No? When they train AI's on data they lose control of that data. If the data is sensitive, they aren't being responsible.
GPT models are as you say dumb statistical models, I agree. But in its weights are encoded ghost images of its training data. The model being dumb is not sufficient to make the data storing itself defensible in my opinion.
Sure, but are you suggesting they somehow encoded, falsely, that they were a murder?
Because it's very unlikely.
It fabricated this from no where. So there's nothing to delete. Because it's just a response to a prompt.
No I'm not, that part is absolutely hallucinated. Where the problem comes in is that it then output correct personal information about him and his children. A to me clear violation of GDPR.
That's not what they're asking for. They're asking for the ability for it to not generate that sentence again.