this post was submitted on 20 Nov 2024
37 points (95.1% liked)

Technology

59472 readers
5179 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] webghost0101@sopuli.xyz 7 points 2 hours ago* (last edited 2 hours ago) (2 children)

I have mixed feelings about this prosecution of ai deepfakes.

Like obviously people should have protection against becoming a victim of such and perpetrators should be held accountable.

But the line “feds are currently testing whether existing laws protecting kids against abuse are enough to shield kids from AI harms” would be a incredibly dangerous precedent because those are mostly designed for actual physical sex crimes.

As wrong as it is to create and distribute ai generated sex imagery involving non consenting people it is not even remotely as bad as actual rape and distributing real photos.

[–] Blueberrydreamer@lemmynsfw.com 5 points 1 hour ago* (last edited 1 hour ago)

I don't think you're on the right track here. There are definitely existing laws in most states regarding 'revenge porn', creating sexual media of minors, Photoshop porn, all kinds of things that are very similar to ai generated deep fakes. In some cases ai deepfakes fall under existing laws, but often they don't. Or, because of how the law is written, they exist in a legal grey area that will be argued in the courts for years.

Nowhere is anyone suggesting that making deepfakes should be prosecuted as rape, that's just complete nonsense. The question is, where do new laws need to be written, or laws need to be updated to make sure ai porn is treated the same as other forms of illegal use of someone's likeness to make porn.

[–] rottingleaf@lemmy.world -2 points 2 hours ago (2 children)

Creating and distributing anything should be legal if no real person suffers during its creation and if it's not intended at defamation, forgery, such things.

[–] AstralPath@lemmy.ca 3 points 1 hour ago (2 children)

Bruh how is creating and distributing a non-consensual nude-ified picture of a young girl not a cause for suffering for the victim? Please, explain that to the class.

Did you just not go to school as a kid? If so, that would explain your absolute ineptitude on this topic. Your opinion is some real "your body, my choice" kind of energy.

[–] krashmo@lemmy.world 1 points 29 minutes ago

There's a legitimate discussion to be had about harm reduction here. You're approaching this topic from an all-or-nothing mindset but there's quite a bit of research indicating that's not really how it works in practice. Specifically as it relates to child pornography the argument goes that not allowing artificial material to be created leads to an increase in production of actual child pornography which obviously means more real children are being harmed than would be if other forms were not controlled in the same fashion. The same sort of logic could be applied to revenge porn, stolen selfies, or whatever else we're calling the kind of thing this article is referring to. It may not be an identical scenario but I still think it would be fair to say that an AI generated image is not as damaging as a real one.

That is not to say that nothing should be done in these situations. I haven't decided what I think the right move is given the options in front of us but I think there's quite a bit more nuance here than your comment would indicate.

[–] rottingleaf@lemmy.world 1 points 44 minutes ago

Read my comment again.

Your opinion is some real “your body, my choice” kind of energy.

My advice to you would be to improve your reading comprehension before judging this way.

In particular, the word "defamation".

[–] Alphane_Moon@lemmy.world 3 points 1 hour ago (2 children)

You would be fine with AI-gen porn images of your teenage daughter being distributed around the internet?

[–] rottingleaf@lemmy.world 2 points 43 minutes ago (1 children)

I take it, the word "defamation" is not part of your lexicon.

[–] Alphane_Moon@lemmy.world 1 points 26 minutes ago (1 children)

The issue being discussed does not fall under defamation.

[–] rottingleaf@lemmy.world -1 points 16 minutes ago

Making forged pics of someone else falls under defamation.

It's very clearly not rape, sexual abuse, child pornography or non-consensual pornography.

[–] droporain@lemmynsfw.com -5 points 1 hour ago (1 children)

Meanwhile in reality check out what she is distributing through Snapchat and only fans... Maybe pursuing the actual crimes first then if there's spare resources go after fiction.

[–] todd_bonzalez@lemm.ee 4 points 1 hour ago (2 children)

Big "but what was she wearing?" energy here.

I don't give a shit if she's doing Shein bikini hauls on Youtube. If you use AI to nudify her pictures, you're manufacturing child pornography, and deserve the full consequences for doing that.

As for OnlyFans, they are quite strict about age requirements. Children aren't running OF accounts. You just hate women and needed to bring up OF to slut-shame.

[–] sunzu2@thebrainbin.org 1 points 41 seconds ago

If you use AI to nudify her pictures, you're manufacturing child pornography, and deserve the full consequences for doing that.

Somebody in non US satellite foreign state can go and do that now from the youtube "bikini hauls" since they publicly avaoialble

What are you or the feds gonna about that, chief?

If that is your or her concern, don't post pictures online. Otherwise, you are literally the mercy of the internet. Privacy 101.

I am sure giving feds extra powers on this won't end like everything else, ie abused against lesser peons.

[–] rottingleaf@lemmy.world 1 points 39 minutes ago (1 children)

If you use AI to nudify her pictures, you’re manufacturing child pornography, and deserve the full consequences for doing that.

No, equating this to an actual child being raped is incorrect. These are not crimes of remotely equal magnitude.

Comparing a person who raped a child, made photos and distributed them to a person who used Photoshop or an AI tool is, other than just evil, reducing the meaning of the former.

[–] Arkouda@lemmy.ca 1 points 2 minutes ago

It is weird how hard you have been defending the production of child pornography in this thread.