this post was submitted on 18 Jan 2024
429 points (95.2% liked)

Technology

59446 readers
4722 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Rep. Joe Morelle, D.-N.Y., appeared with a New Jersey high school victim of nonconsensual sexually explicit deepfakes to discuss a bill stalled in the House.

you are viewing a single comment's thread
view the rest of the comments
[–] aesthelete@lemmy.world 2 points 10 months ago* (last edited 10 months ago) (6 children)

And I don't think you're fully understanding that the above is some type of fantasy you have, and will not actually be what the future is like at all.

[–] curiousaur@reddthat.com 0 points 10 months ago (5 children)

It's probably a bit of an exaggeration, but my point stands. It's going to be so easy for anyone to see ai gen material of anyone else, no one is going to care anymore.

[–] aesthelete@lemmy.world 2 points 10 months ago* (last edited 10 months ago) (4 children)

I don't even think that's necessarily true. If you make it illegal and/or platforms ban it, you're already taking a step toward making it more difficult to do.

I think throughout this thread you're mistaking the technically possible for the probable or likely.

By making it illegal, you essentially eliminate the commercial incentive for making it easy. Every barrier to doing something makes it more unlikely that people will do it. I understand that there is an inherent motive for people to do it anyway, but, every hoop they have to jump through (e.g. setting up their "own, local AI") reduces the likelihood of them doing it.

People don't even run their own email servers, music servers, video servers, etc. etc. etc....most people don't even "jail break" devices...many don't even store a local cache of regular porn...why the hell would most people bother themselves with setting up a local generative AI instance for this purpose?

Outlawing it and banning it from platforms makes it much more within the realm of the creepy basement weirdo rather than something that is as inevitably ubiquitous as you're saying it will be.

Policy is very often about reduction of harms rather than elimination of harms. It's not the black and white realm that you're trying to make it out to be.

[–] curiousaur@reddthat.com 1 points 10 months ago (1 children)

It's not illegal to to work on, sell, or distribute the models. And making that illegal is what the first commenter said would be dangerous to do, since then regular people wouldn't be able to compete with corporation's abilities.

Once the models and portable hardware are good enough, and it's just a matter of time, I think you're underestimating how ubiquitous it will become.

Every teenage boy will have a pair of nudie glasses in the form of their smartphone running open source models, and you think they're just going to not use them?

[–] aesthelete@lemmy.world 2 points 10 months ago* (last edited 10 months ago) (1 children)

I think you again vastly overestimate how many people are going to run their own AI versus using a sanitized, policy-driven, managed platform version that's cloud based (e.g. Dall-E and ChatGPT right now).

It's possible today (and usually better) to do a lot of things locally, but yet still almost everything routes through an app to a platform on your smartphone and the few remaining things that don't route through a platform using your phone's browser.

[–] curiousaur@reddthat.com 1 points 10 months ago (1 children)

When it becomes one click to see the chick across from you naked, tell me how many 16 year old boys won't. You are far too naive to be having this conversation.

[–] aesthelete@lemmy.world 1 points 10 months ago* (last edited 10 months ago)

It's not naive to think that corporations will continue to win the "AI" war. It's actually pretty naive to think otherwise.

I also dunno why you think that all of the resources in oss AI will focus their efforts on making it easy to generate excellent, likely already illegal deep fake porn of random teenagers in "one click".

I've been using oss for decades and almost nothing is that easy to do even when it could be. Why would people focus their efforts on this?

Also also, I don't get why you think that generating AI porn of people around you is:

A) so much better than just watching the millions of hours of already available porn

B) anything even remotely similar to "seeing someone naked"

load more comments (2 replies)
load more comments (2 replies)
load more comments (2 replies)