this post was submitted on 02 Mar 2024
203 points (89.2% liked)
Technology
59219 readers
3314 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Read a bit of the court filing, not the whole thing though since you get the gist pretty early on. Jornos put spin on everything, so here's my understanding of the argument:
It's honestly not the worst argument.
I actually agree with this. This technology should be open. I know that there are arguments to keep it closed, like it could be misused, etc. However, I think that all the scary stories about AI are also a way to keep attention away from the fact that if you have a monopoly on it, you have enormous power. This power will grow when the tech is used more and more. If all this power is in the hands of a commercial business (even though they say they aren't), then you know AI is going to be misused to gain money. We do not have clear insight in what they are doing and we have no reason to trust them.
You also know that bad actors, like dictatorial governments will eventually get or develop the technology themselves. So, keeping it closed is not a good way to protect it from that happening. At the same time, you are also keeping it from researchers who could investigate how to use and develop it further to be used responsibly and to the benefit of humanity.
Also, they relied on data generated by people in society who never got any payment or anything for that. So, it is immoral to not share the results with that same people in society openly and instead keeping it closed. I know they used some of my papers. However, I am not allowed to study their model. Seems unfair.
The dangers of AI should be kept at bay using regulation and enforcement by democratically chosen governments, not by commercial businesses or other non-democratic organisations.
I don't want Musk to be right, but I have to admit, it sounds legit.
Yeah, fuck "it's not in the terms of a contract". It's fraud.
You can't advertise yourself as a nonprofit organization for the public good, collect donations under that pretense, then just privatize anything you learn for profit.
People don't donate to for profit companies.