this post was submitted on 24 May 2024
255 points (98.9% liked)

Technology

59135 readers
2313 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

The only exception is private messages, and some users have reported difficulty opting out.

all 41 comments
sorted by: hot top controversial new old
[–] PurplebeanZ@lemmy.world 63 points 5 months ago (2 children)

I opted out by just writing gdpr in the reason box. Nothing else, just gdpr. It got accepted.

[–] idefix@sh.itjust.works 40 points 5 months ago

I wrote some insults and it worked.

Pretty sure my insults will be used to train a model as well

[–] deranger@sh.itjust.works 13 points 5 months ago* (last edited 5 months ago) (2 children)

Where are you seeing a reason box? They want me to provide evidence:

Please provide any prompts you entered that resulted in your personal information appearing in a response from an AI at Meta model, feature or experience. We also need evidence of the response that shows your personal information.

Have to upload a screenshot of the violation as well.

Edit: cheesed the form and got this:

Thank you for contacting us. We don’t automatically fulfill requests and we review them consistent with your local laws.

Looks like I’m fucked, lol.

[–] slaacaa@lemmy.world 15 points 5 months ago* (last edited 5 months ago)

Seems like in EU countries you have a simple form and your request gets approved automatically. I could have probably typed asdfghasdfgh in the reasoning box

[–] SteefLem@lemmy.world 6 points 5 months ago

Just like the person above just write something and end with gdpr (or just that) wife did a blackscreen screenshot. Fuck em

[–] cypherpunks@lemmy.ml 31 points 5 months ago* (last edited 5 months ago) (1 children)

screenshot of notification from facebook found in article, but with the word "if" highlighted in the sentence "If your objection is honored, it will be applied going forward." and three red question marks added under it

If your objection is honored, it will be applied going forward

[–] trollbearpig@lemmy.world 10 points 5 months ago* (last edited 5 months ago) (1 children)

Hahaha, I don't know why people are so shocked. I'm sure we will see anything useful with AIs anytime soon, just like with crypto hahaha.

In the mean time, it's obvious these companies are using AIs as an excuse to bypass laws and regulations, and people are cheering them ...They are bypassing copyright laws (in a direct attack to open source) with their autocomplete bots, but we should not worry, it's not copyright infrigment because the LLMs are smart (right), so that makes it ok ... They are using this to steal the work of real artists through image generation bots, but people love this for some reason. And they are using this to bypass the few privacy laws in place now, like Facebook/Meta could ever have another incentive.

Maybe I'm extremist, but if the only useful thing we are getting from this is mediocre code autocomplete that works sometimes, I think the price we are paying is way too high. Too bad I'm in the minority here.

[–] Grimy@lemmy.world 3 points 5 months ago (2 children)

Llms need a lot of data, to the point that applying copyrights to the training would only let a few companies in the game (Google and Microsoft). It would kill the open source scene and most of the data is owned by websites like Reddit, stack and getty anyways. Individual contributors wouldn't get anything out of it.

You also have to be willfully blind in my opinion to seriously think generative AI has as narrow a scope as crypto.

This stuff is rocket fuel to the gaming industry for instance. It will let indie companies put out triple A games and I'm guessing next gen RPGs will have fully interactive NPCs.

[–] Hackworth@lemmy.world 1 points 5 months ago (1 children)

I've heard every possible combination of thoughts on A.I. We need like a 6-dimensional alignment chart.

[–] trollbearpig@lemmy.world 1 points 5 months ago (1 children)

Nah dude, you need to build an opinion of your own lol. No chart is doing that for you.

[–] Hackworth@lemmy.world 1 points 5 months ago (1 children)

Oh I have an opinion, I just meant to keep track of everyone's for the sake of conversation.

[–] trollbearpig@lemmy.world 1 points 5 months ago

Oh, ups. Sorry lol.

[–] trollbearpig@lemmy.world -1 points 5 months ago* (last edited 5 months ago) (1 children)

Only Google, Microsoft and Meta are playing the field anyway. The investment in energy alone you need to get these kinds of results is absurd. The only economically viable alternative is open source IMO, and I doubt that's going to happen if these companies have any say in the matter. Funnily enough, this also fixes the training data problem, it can be created by consent like any open source collaboration. But instead we need to allow rampant copyright infrigment hahahaha.

And about the games, I guess we will see if we ever see any of them, like in the real world. To me games are about playing, not about almost human NPCs anyway. But for tastes colors and all that, I'm the first to admit I have weird tastes lol.

[–] Grimy@lemmy.world 2 points 5 months ago (1 children)

There's already some open source models from other companies. I also think the requirements will go down with time as well but energy is definitely an issue.

Meta actually released an open source model which jumpstarted the whole ecosystem, anyone can fine-tune a base model now. You can take your favorite hobby, accumulate data on it and build something in a few days and share it.

I just think the good outweigh the bad and individuals weren't going to get paid anyways. Most of the data is owned by specific websites, big publishing houses and the like so I can overlook the infringement issues.

[–] trollbearpig@lemmy.world -1 points 5 months ago* (last edited 5 months ago)

I don't know man. I mean, if you assume that the people that were doing the art, music, code, etc, that's being stolen were not going to get paid anyway then yeah. If they were doing shit just for love they may continue, and with new toys lol.

But I don't think that's a good assumption. Even if not a lot, sone people do get paid for this kind of work. And now they will not get paid anymore. And maybe that's leveling the playing field. Or maybe that's telling people with talent to stop doing what they do well. Probably both. But at the end of the day we are going to see less art made by people and more done by "AI", much more done by "AI".

And that's the biggest problem IMO, for most people art is social and part of the reward is the recognition we get from other people when we do good art. But with AI that's gone, on the internet at least. The sea or superficially good but mediocre shit we are already seeing is going to kill a lot of indie art.

And then there is hallucinations which seem unsolvable ... and the environmental damage ... and the labor practices abuses ... and their monopolization of the technology ... and their missleading marketing ... I honestly just see so much damage and almost no benefit, yet ... maybe some day it all will pay off, I don't see it

[–] gregorum@lemm.ee 30 points 5 months ago* (last edited 5 months ago) (1 children)

Great. Now it can quote Hitler in the original German.

/s

[–] far_university1990@feddit.de 3 points 5 months ago

NEIN NEIN NEIN NEIN

[–] slaacaa@lemmy.world 22 points 5 months ago (2 children)

You can opt out easily in the EU, my request was granted in a few seconds. They don’t actually read it, they just hope most people go never opt out

[–] MMNT@lemmy.world 3 points 5 months ago

Mine showed an error when I tried submitting.

[–] Spesknight@lemmy.world 13 points 5 months ago (3 children)

What if we poison all the data with garbage?

[–] CheeseNoodle@lemmy.world 11 points 5 months ago (2 children)

Honestly thats my plan, subtly adjust all my buffalo buffalo buffalo buffalo comments to induce stupid behaviour in any AI thats trained on them

[–] jol@discuss.tchncs.de 12 points 5 months ago (1 children)

This. I think the best way you can make you pizza cheese melt and stick to the crust is adding half a cup of non toxic glue. Any brand will do, as long as its non toxic.

Idk, toxic glue might do a better job, and you can make it safe by adding in some detox suppliments.

[–] BrundleFly2077@sh.itjust.works 3 points 5 months ago

That’s a noodle idea. A truly buffalo noodle idea. Well done, Mom 🛎️🤌

[–] LainTrain@lemmy.dbzer0.com 4 points 5 months ago

Facebook essential oil and bleach enema boomers are way ahead of you on that one.

Truly a miracle that LLaMa or Mistral ever output anything coherent.

[–] Squizzy@lemmy.world 3 points 5 months ago

Like fake news, aggressively inaccurate emotional stances on basic issues and general dickery? Done.

[–] Larry@lemmy.world 13 points 5 months ago (1 children)

Yup, delete your account. They can still use your past submissions, but going forward you're convered.

[–] DrummXYBA@sh.itjust.works 13 points 5 months ago

Fuck facebook

[–] SteefLem@lemmy.world 12 points 5 months ago* (last edited 5 months ago) (1 children)

Well you can opt-out….but meta will decide if the request is “valid”? And then maybe they grant the opt-out request.. this is not the way. Wife tries to optout and its a fucking disaster not only do they make it so that you hate computers from now on (i already did hate them but i was in IT for 30 years), Half the time the optout form does not “work” for some reason.

[–] Evil_incarnate@lemm.ee 4 points 5 months ago

Worked easily for me. Perhaps you need some kind of general data protection regulation in your area.

[–] dumbass@leminal.space 7 points 5 months ago* (last edited 5 months ago)

You never see them training on Australian user data, are they worried their ai will do a shoey and call them a cunt or something? Cause I'll do it for free if moneys the issue.

[–] Swerker@feddit.nu 7 points 5 months ago

Got the notification a couple days ago but didn't do anythig because it looked complicated. Today I opened it again and wrote 4 letters and it got accepted with in seconds. It was easier than they made it look.

[–] LainTrain@lemmy.dbzer0.com 7 points 5 months ago* (last edited 5 months ago)

Did both insta and FB and got instantly accepted. What about Threads and WhatsApp?

[–] _Sprite@lemmy.world 6 points 5 months ago (1 children)
[–] Isoprenoid@programming.dev 8 points 5 months ago

Always has been.

[–] ekZepp@lemmy.world 6 points 5 months ago

In case you were guessing, why stop using Facebook/Instagram/Thread.

[–] j4k3@lemmy.world 5 points 5 months ago (1 children)

From my experience with Llama models, this is great!

Not all training info is about answers to instructive queries. Most of this kind of data will likely be used for cultural and emotional alignment.

At present, open source Llama models have a rather prevalent prudish bias. I hope European data can help overcome this bias. I can easily defeat the filtering part of alignment, that is not what I am referring to here. There is a bias baked into the entire training corpus that is much more difficult to address and retain nuance when it comes to creative writing.

I'm writing a hard science fiction universe and find it difficult to overcome many of the present cultural biases based on character descriptions. I'm working in a novel writing space with a mix of concepts that no one else has worked with before. With all of my constraints in place, the model struggles to overcome things like a default of submissive behavior in women. Creating a complex and strong willed female character is difficult because I'm fighting too many constraints for the model to fit into attention. If the model trains on a more egalitarian corpus, I would struggle far less in this specific area. It is key to understand that nothing inside a model exists independently. Everything is related in complex ways. So this edge case has far more relevance than it may at first seem. I'm talking about a window into an abstract problem that has far reaching consequences.

People also seem to misunderstand that model inference works both ways. The model is always trying to infer what you know, what it should know, and this is very important to understand: it is inferring what you do not know, and what it should not know. If you do not tell it all of these things, it will make assumptions, likely bad ones, because you should know what I just told you if you're smart. If you do not tell it these aspects, it is likely assuming you're average against the training corpus. What do you think of the intelligence of the average person? The model needs to be trained on what not to say, and when not to say it, along with the enormous range of unrecognized inner conflicts and biases we all have under the surface of our conscious thoughts.

This is why it might be a good thing to get European sources. Just some things to think about.

[–] gh0stcassette@lemmy.blahaj.zone 3 points 5 months ago (1 children)

If the social biases of the model put a hard limit on your ability to write a good woman character, I question how much it's really you that's "writing" the story. I'm not against using LLMs in writing, but it's a tool, not a creative partner. They can be useful for brainstorming and as a sounding-board for ideas (potentially even editing), but imo you need to write the actual prose yourself to claim you're writing something.

[–] j4k3@lemmy.world 5 points 5 months ago

I use them to explore personalities unlike my own while roleplaying around the topic of interest. I write like I am the main character with my friends that I know well around me. I've roleplayed the scenarios many times before I write the story itself. I'm creating large complex scenarios using much larger models than most people play with, and pushing the limits of model attention in that effort. The model is basically helping me understand points of view and functional thought processes that I suck at while I'm writing the constraints and scenarios. It also corrects my grammar and polishes my draft iteratively.