this post was submitted on 28 Sep 2023
124 points (84.8% liked)

No Stupid Questions

35888 readers
1724 users here now

No such thing. Ask away!

!nostupidquestions is a community dedicated to being helpful and answering each others' questions on various topics.

The rules for posting and commenting, besides the rules defined here for lemmy.world, are as follows:

Rules (interactive)


Rule 1- All posts must be legitimate questions. All post titles must include a question.

All posts must be legitimate questions, and all post titles must include a question. Questions that are joke or trolling questions, memes, song lyrics as title, etc. are not allowed here. See Rule 6 for all exceptions.



Rule 2- Your question subject cannot be illegal or NSFW material.

Your question subject cannot be illegal or NSFW material. You will be warned first, banned second.



Rule 3- Do not seek mental, medical and professional help here.

Do not seek mental, medical and professional help here. Breaking this rule will not get you or your post removed, but it will put you at risk, and possibly in danger.



Rule 4- No self promotion or upvote-farming of any kind.

That's it.



Rule 5- No baiting or sealioning or promoting an agenda.

Questions which, instead of being of an innocuous nature, are specifically intended (based on reports and in the opinion of our crack moderation team) to bait users into ideological wars on charged political topics will be removed and the authors warned - or banned - depending on severity.



Rule 6- Regarding META posts and joke questions.

Provided it is about the community itself, you may post non-question posts using the [META] tag on your post title.

On fridays, you are allowed to post meme and troll questions, on the condition that it's in text format only, and conforms with our other rules. These posts MUST include the [NSQ Friday] tag in their title.

If you post a serious question on friday and are looking only for legitimate answers, then please include the [Serious] tag on your post. Irrelevant replies will then be removed by moderators.



Rule 7- You can't intentionally annoy, mock, or harass other members.

If you intentionally annoy, mock, harass, or discriminate against any individual member, you will be removed.

Likewise, if you are a member, sympathiser or a resemblant of a movement that is known to largely hate, mock, discriminate against, and/or want to take lives of a group of people, and you were provably vocal about your hate, then you will be banned on sight.



Rule 8- All comments should try to stay relevant to their parent content.



Rule 9- Reposts from other platforms are not allowed.

Let everyone have their own content.



Rule 10- Majority of bots aren't allowed to participate here.



Credits

Our breathtaking icon was bestowed upon us by @Cevilia!

The greatest banner of all time: by @TheOneWithTheHair!

founded 1 year ago
MODERATORS
 

There has been a ton of CSAM and CP arrests in the US lately, especially from cops and teachers, along with at least one female teacher seducing boys as young as 12. I cannot understand the attraction to kids. Even teens. Do these people think they are having a relationship, or it is somehow okay to take away another human beings' innocence? Is it about sex, power, or WTH is it? Should AI generated CSAM and CP be treated the same as a real person since it promotes the same issues? I am a grandfather, and I am worried about how far this will go with the new AI being able to put anyone's face into a Porno movie too.

It seems to me that a whole new set of worldwide guidelines and laws need to be put into effect asap.

How difficult would it be for AI photo apps to filter out words, so someone cannot make anyone naked?

top 50 comments
sorted by: hot top controversial new old
[–] Veraticus@lib.lgbt 163 points 1 year ago* (last edited 1 year ago) (6 children)

I'm only going to answer the first part of your question, not the AI/generated part.

No one really chooses what or who they're attracted to; it kind of just happens to you. For example, you might be watching a TV show and someone gets lightly, comically spanked... and suddenly a light bulb goes off above your head and you think, "whoa, that might actually be kinda fun." People are wired in ways we don't understand to want things we don't even know we want.

To that extent, pedophiles are themselves victims of their own desires; there's no "logic" behind it. It's simply an urge they experience.

Of course that doesn't make succumbing to this urge excusable, and any children who are impacted are of course victims and the pedophiles, predators. But no one is training pedophiles in pedophile camp. It's just humans being human, unfortunately.

[–] HappycamperNZ@lemmy.world 75 points 1 year ago (1 children)

This is something many people fail to realize - while society hates it exists, it is just an urge the same as my desire for women. We have just grown as a society to say this isn't right (correctly). There are many who have the urge, and don't follow up on it but its still there and they are a victim as well.

Fully agree though, this does not excuse those who act upon it, promote it or sell it.

[–] CaptainEffort@sh.itjust.works 29 points 1 year ago (2 children)

Unfortunately I think it’s probably in the same vein as any fetish or preference, so completely out of their control.

Obviously people who act on it are the scum of the earth, but those who simply battle with the urge I have nothing but sympathy for. I can’t even imagine how horrible it is to have to deal with that daily and never be able to do anything about it, or even really talk to anyone about it.

load more comments (2 replies)
[–] Fredselfish@lemmy.world 51 points 1 year ago (3 children)

I have heard that kids that molest kids or end up fooling around a super young age can make them grow up wired to be attracted to young kids or teens.

The real issue no one wants to address is people who have these desires and know it wrong have nowhere to turn to for help.

Even If they haven't abuse anyone come out tell a therapist or someone you are attracted to kids will and probably can get you locked up. There are those who never affend but a lot do because either a) they accept what they are and have no moral objections to it or b) can't get the help needed to fight the urges and end up offending.

As @Veraticus said there is no easy answer because it's not a choice. Be like asking you why you like women or why people are gay. They wired that and unfortunately I don't think you can cure it.

We definitely need to address access to any kind porn of it and if someone offends we must lock them away for their own good. Not saying prison but somewhere they can be mentally evaluated.

[–] Instigate@aussie.zone 25 points 1 year ago (1 children)

There is definitely a link between having experienced sexual abuse as a child without any therapy or counselling to help them make sense of it and then later on sexually abusing other children, but it’s not super clear-cut and definitely not predictable.

[–] Fredselfish@lemmy.world 13 points 1 year ago

Yes same with girls who are raped or molested become promiscuous but doesn't mean all girls in that situation will. Definitely why we need better sex education in America so we can teach kids the signs if an adult being inappropriate and learn about their own bodies.

load more comments (2 replies)
[–] XbSuper@lemmy.world 6 points 1 year ago

This is one of the most sane responses I've ever seen.

I am one of those poor souls who has these urges, but has never, and will never, act on them.

I'm willing to open myself to an AMA for anyone interested.

[–] DingoBilly@lemmy.world 5 points 1 year ago (3 children)

This is the most accurate answer, and the fact it's all cultural/social is quite important as well.

If you were born a few thousand years ago it may be completely reasonable to sleep with a kid. Hell the kid is probably your slave so you could literally do whatever you want with them.

But just as I don't understand certain fetishes or even just people attracted to the same sex, others won't understand why people would be attracted to kids.

[–] HelixDab2@lemm.ee 3 points 1 year ago (1 children)

If you were born a few thousand years ago it may be completely reasonable to sleep with a kid.

That simply is not correct. While kids have been having sex with kids since, like, forever, looking back to medieval Europe, or even Rome, it was not normal for adults to be having sex with children, especially prepubescent children. Marriage ages for people that weren't nobility (e.g., getting married off to solidify political alliances) were typically early 20s. Children getting pregnant has always been a very dangerous proposition, since it's much more likely to lead to maternal mortality.

Part of the confusion about all of this is that we think people used to live compressed lives, since we see average lifespans in the 40s. But that's not taking into account the ferocious infant mortality rates; if you survived into your late teens, you were probably going to make it to your late 60s/early 70s, even in the middle ages. We think that, since 'average' life spans were 40-odd years, that people must have been marrying very young, but the evidence doesn't really bear that out.

load more comments (1 replies)
load more comments (2 replies)
load more comments (2 replies)
[–] Deestan@lemmy.world 39 points 1 year ago* (last edited 1 year ago)

To answer some of the questions:

I cannot understand the attraction to kids.

There was a TV interview with people who were seeking help for pedophilia. They described it as just plain horny sexual attraction that they knew they had to not act on. I guess people have different reasons, and some probably manage to rationalize it as "relationships" as you say.

Should AI generated CSAM and CP be treated the same as a real person since it promotes the same issues?

Whether it is a modified image of a real person, or a pure generated picture, they will fall under the same laws as for depicting it which is already uncontoversially illegal.

How difficult would it be for AI photo apps to filter out words, so someone cannot make anyone naked?

Hard, as there are many ways to describe nudity or encourage the generator to weigh towards nudity. "Person with visible thighs, no skirt" and such.

Easier to leave nudity out of the training data, which is already common.

Then hard again because anyone can throw together a new image generator trained on what they want and no word filters.

[–] dameoutlaw@lemmy.ml 30 points 1 year ago (1 children)

I see people here attempting to equate it with a natural attraction and or a fetish. We all have internet access and can look it up. It is in the DSM-5 and ICD-10 both pedophilia and pedophilic disorder. Especially, in this more advanced age of medicine, science and society. If it was natural I believe the corrections would’ve been made and or strongly advocated for, it’s not. What is advocated is using terminology correctly, encouraging those that experience this to feel comfortable to tell their truth and seek help. I believe some of you guys comments are very dangerous and some of the upvotes and downvotes are concerning and makes it difficult to distinguish if you are in support of protecting children. The point is please don’t just blanket label it and compare it to things that aren’t harmful, illegal and consensual.

load more comments (1 replies)
[–] VelvetStorm@lemmy.world 28 points 1 year ago* (last edited 1 year ago) (1 children)

Edit: also don't be sexist, call it what it is rape, not seduction. Just because it was a woman doesn't make it not rape. Calling it anything else is doing a disservice to all of the male victims of female on male rape.

If the AI porn is depicting real life underage people then it should be and is(in some places) illegal. Now I don't like it and find it reprehensible but if it's not depicting real-life people then it should be legal.

I think if you are into that then you should be able to seek professional help without a fear of it ruining your professional and personal life, but if you are attracted to kids then it is your moral responsibility and obligation to not work with or be around kids.

This is a mental issue and it should be treated like one and we should be trying to understand it and find ways to prevent and treat it.

[–] snausagesinablanket@lemmy.world 10 points 1 year ago* (last edited 1 year ago)

I was reading this, and it made me remember how a dude in Australia I believe bought an underage sex doll. It ended up being flagged somehow, and the government arrested the guy when it arrived. I have no idea what happened to the guy.

Was this guy trying to control his urges by using this hunk of rubber, or is that a crime too? This is very edgy, and I am thinking some countries won't bother with them, while others might incur the death penalty.

[–] HelixDab2@lemm.ee 25 points 1 year ago (2 children)

There are multiple parts to your question. I'm going to try to break it down.

First, there's a difference between a pedophile and a child molester. Pedophilia is a sexual attraction to children, but it does not, by itself, require the person to take action. A child molester is a person that sexually assaults children. It's the difference between being heterosexual, and being a rapist; you can be straight and still be entirely celibate.

Child molesters may not be sexually attracted to children at all; some might be, but people that commit rape aren't usually doing it solely for sexual gratification, although sex is definitely part of it.

We don't know how common pedophilia is because of how heavily stigmatized it is.

You don't understand how a person could be sexually attracted to children; the simplest way to explain it is to ask if you can understand how a man can be sexually attracted to another man. IIRC, most research indicates that pedophilia probably is a sexual orientation, much like being straight, gay, or bisexual is (except that there is no moral or ethical way for a pedophile to have a sexual or romantic relationship with a child; that is always both predatory and criminal). Do pedophile child molesters believe that they're having a relationship? Some of them, yes. They're able to delude themselves into believing that the child wants the attention and sex (really sexual assault), when they're--probably--the one that has groomed the child in the first place.

I cannot understand the attraction to kids. Even teens.

I can. When I was a child, I was sexually attracted to my peers. 14yo kids are having sex with each other, so clearly they're attracted to each other. As an adult, I can see women in their 20s as being sexually attractive, while still having zero interest in them (y'all seem really young, and not in a good way, if y'know what I mean). Sexual maturity isn't a magical thing that happens when you hit 18 (or whatever the age of consent is where you live); it's a sliding scale.

Should AI generated CSAM and CP be treated the same as a real person since it promotes the same issues?

I don't think that you can make a person into a pedophile, any more than you can make a person gay. A person either is, or isn't, a pedophile, and CG CSAM isn't going to change that. So the question is, does CG CSAM make it more likely that a pedophile will end up sexually abusing a child? My intuition says that it will not, in the same way that the proliferation of pornography has not made sexual assault of adults more common. (Some research indicates that the availability of pornography has decreased rates of sexual assault.) Child pornography is illegal--in part--because it cannot be produced without causing real harm to children. CG CSAM doesn't cause real harm to any person though; unless there's evidence that it increases the rates of child sexual abuse, I don't think that the squick factor is a reasonable basis for banning it. OTOH, adult pornography has generally led to a relaxation of sexual mores and norms--which I believe is a generally positive thing--and it's possible that CG CSAM would normalize child sexual abuse sufficiently that libertarians would be able to severely weaken age of consent and statutory rape laws. I don't really know, TBH; I'd want to see more research rather than reflexively banning it.

[–] Chailles@lemmy.world 2 points 1 year ago

I don’t think that you can make a person into a pedophile, any more than you can make a person gay.

If you really think about it, we've seen arguments like that before. That pornography creates rapists. That violent video games creates murderers. And that's just strictly on the consumption of media.

[–] MystikIncarnate@lemmy.ca 2 points 1 year ago (1 children)

Articulate and concise. I like it.

I don't really have anything more to add, except that we as a society need to make up our collective minds on cg and AI csam/CP soon.

I definitely think that cg/AI content is less bad (still bad, but less so) because there's no harm being done to real children from it; but those AI image engines needed to be trained on some form of content to be able to generate the images that they do, so I'm not sure how the training images factor in, or what would even be used for training that AI... I know rather little about how AI is trained at the moment, so I'm not sure if it can be done without source csam material or not... IMO, that factors into the morality of the output.

I definitely agree that sexual preference toward minors (aka paedophilia) is just that, a sexual preference; and that, in and of itself does not make someone a sex offender in the same way that being heterosexual doesn't make you a rapist or other form of sexual "deviant" (or however you want to say that. It's interesting to me to think that paedophiles may have a semi-legal way of getting porn for themselves (which causes no harm to children). I feel a bit bad for paedophiles in that they're basically forced to have relationships with persons that they don't find very sexually attractive, else they break the law. Not bad enough that I think that laws should change our anything, it's just a crap situation. It would be like having a preference towards men, as a man, in a world of heteros. The men are there and you're interested in them, but none of them are interested in you. Almost always that's not the case, there's other homosexual men that exist, no matter how rarely... in the case with pedos, there are exactly zero underage people who they can interact with at all sexually. I still don't think that should change, but at least with the internet, a gay man can go and find porn that interests them. With pedos is literally a crime to even look at, possess, or make any porn that appeals to them.

I can sympathize with the impossibility of their situation, that's all. For the record, I'm just done cis male with no interest in anyone too young to date. I can recognise their attractive qualities without being attracted to them (speaking mostly about those that have reached their sexual maturity here, who are still not 18 or whatever)... I can understand it, I'm not so hateful to want anyone who feels attraction to young people to die or anything, but young people don't have the experience to understand the situation they're getting into, when they're being mislead or gas lit, etc (though to be fair, a lot of "mature people seem to not know either, but that's another discussion)... Fact is, they're shit out of luck.

I'm sure many are forced into celibacy just to be lawful. I don't think any grown adult wants to be forced to be celibate; so I can understand the plight. AI/cg porn, tailored to that specific preference may give pedos an outlet that they can utilize to temper their urges and keep them on the right side of the law here. Of course it won't solve the problem entirely, the same way that rapists are still a thing, but it may severely reduce illegal activity and harm to children.

But I agree, it's a slippery slope (so to speak) because it can easily evolve into lowering the age of consent, and bringing back child marriages and such. Which IMO, isn't a desired outcome. I also don't think that content should intermingle with either social networks or existing porn sites, since it's so specific, it should be relegated to specific sites and not left flapping around the internet. It's also a vast minority of people that are afflicted, so segregation may be a minimum measure to keep things somewhat clean. I know I don't want AI generated CP content mixed in with my usual porn browsing... I'm sure there's plenty of people in the same boat, so IMO that's a minimum. But I'm only one voice in the society, so I don't make the decision; I'm interested to see what decision is finally made and implemented, whenever we get there.

As a disclaimer: I'm not attracted to underage people. I'm also not a doctor or scientist, or psychologist or anything else. I'm not in favor of anything here, besides society making a decision, and I'm just positing that it could be beneficial to society as a whole. I welcome other opinions, except those by people whom are heavily religious. Good day.

[–] HelixDab2@lemm.ee 2 points 1 year ago (1 children)

I feel a bit bad for paedophiles in that they’re basically forced to have relationships with persons that they don’t find very sexually attractive, else they break the law.

I've seen a paper--which I unfortunately did not bookmark--that seemed to indicate that most pedophiles were not exclusively pedophiles; many are able to have romantic and sexual relationships with age-appropriate partners. They also tend to have a distinct gender preference for minors (e.g., a person that is a heterosexual and a pedophile will prefer minors that are in-line with their sexual orientation). The ones that are 'pure' pedophiles--not sexually attracted to any adults at all--do not seem to have a gender preference, which kinda makes sense when you consider secondary sex characteristics as markers of physical maturity, e.g., young boys and girls look physically very similar aside from the genitals themselves. Again - I don't have the reference on this saved, so I might be misremembering, or misrepresenting it, but this is what I recall.

there are exactly zero underage people who they can interact with at all sexually.

There's a genetic disorder--I believe exclusively in women--where they don't 'grow up'; they don't get very tall, they're largely lacking in secondary sexual characteristics, and I believe that they're infertile. I ran into a woman like that--who was with her partner--at a fetish event. It really gave me whiplash, because at first, second, and third glances she looked like she was 12, at an event that had explicit sexual activity in the open. It took a closer look at her face to realize that she was in her 30s. God bless her, she found someone that was attracted to her, and into the same shit she was into. So, y'know, there's that.

[–] MystikIncarnate@lemmy.ca 2 points 1 year ago (2 children)

I've heard of that genetic condition. It's fascinating, and as far as I know, extremely rare.

I know that at least one has spoken publicly about her experience, and they touched on dating and the implication was that most of the people that are interested, are paedophiles, and that didn't sit well for her, and I expect that wouldn't sit well for most people, especially those with that condition.

Fascinating information all around. I don't have a doubt that is accurate.

load more comments (2 replies)
[–] cooopsspace@infosec.pub 21 points 1 year ago* (last edited 1 year ago)

At the end of the day, art is just pixels on a flat surface. Determining whether a depicted individual is under age where it's not obvious sets a dangerous precedent. Is the picture 17 or 18? Who knows.

But the problem is that people have been sexualising people like Emma Watson since she first appeared on screen. That's not okay and rather than sending AI art underground I think society needs to change to normalise education about sex, reproduction and genitalia and address the social issues to treat pedophilia like the disease that it is.

Meanwhile pedophile names are being written about publicly, risking mob violence and further isolation. Not to mention in the US theres a lot of negative attention being put on women's reproduction, childrens sex ed and genitalia and a push to make the whole lot illegal and taboo. Not to mention people teaching their kids pet names for their parts, "uncle Ben touched my heehaw" sounds a lot different to "uncle Ben touched my penis".

Society is a problem, the US particularly is going in the wrong direction on many aspects of sex education.

[–] Astroturfed@lemmy.world 20 points 1 year ago

Don't try to understand it. You aren't going to get a good answer. It's a horrible mental illness level of sexual preference.

Anything can be sexualized with enough impulse and experiences. Everyone's got some weird dark fetish shit. Some of it's illegal in practice. Normal people bury that shit or only discuss it in therapy. While talking it out so they can hopefully never think about it again.

I'm sure there's different answers to this just like "why are there serial killers?". Just be glad it confuses you.

[–] scarabic@lemmy.world 13 points 1 year ago

I imagine there’s some kind of vampiric quality to it. Kids are full of youth and innocence: things we are all constantly losing to time. Especially if someone’s own childhood was robbed from them, I think they will carry around a void they desperately want to fill but never can, because of course abusing a child doesn’t bring these things back to you. Still, many child abuse victims go on to abuse other children later in life, and this may be their drive: to seek the thing they lost. It’s beyond sad. Abusing children is straight up disgusting and terrible, but the convoluted desperation that causes people to do it is truly horrifying in a stranger-than-any-fiction kind of way.

[–] Apepollo11@lemmy.world 12 points 1 year ago* (last edited 1 year ago) (1 children)

I'm only going to tackle the tech side of this...

How difficult would it be for AI photo apps to filter out words, so someone cannot make anyone naked?

Easy. The most popular apps all filter for keywords, and I know that at least some then check the output against certain blacklisted criteria to make sure it hasn't let something slip through.

But...

Anyone can host their own version and disable these features, allowing them to generate whatever they want, in the exactly same way that anyone can write their own story containing whatever they want. All you need is the determination to do it, and some modicum of ability.

People have been been creating dodgy doctored photos long before computers. When Photoshop came out, it became easier, and with AI it's easier still. The current laws about creating and distributing indecent images still apply to these new images though.

[–] Adalast@lemmy.world 2 points 1 year ago (2 children)

Technically the diffusers all have the ability to filter any material from the actual outputs using a secondary CLIP analysis to see if it kicks out any keywords which indicate that a topic is in the image. From what I have seen, most AI generation sites use this method as it is more reliable for picking up on naughty outputs than prompt analysis. AI's are horny, I play with it a lot. All you have to do is generate a woman on the beach and about 20% of them will be at least topless. Now, "woman on the beach" should not he flagged as inappropriate, and I don't believe the outputs should either because our demonization of the female nipple is an asinine holdover from a bunch of religious outcasts from Europe who were chased our for being TOO restrictive and prudish, but alas, we are stuck with it.

load more comments (2 replies)
[–] ThisIsAManWhoKnowsHowToGling@lemmy.dbzer0.com 9 points 1 year ago (1 children)

There are two parts to this problem.

For kids who haven't hit puberty, there is a diagnosable pedophilia disorder. This is mostly genetics. (I'm pretty sure I've met an alpaca that was a pedophile once.) The molester's brain is wired wrong. Nothing to do about that. IMHO, they deserve pity as long as they keep their hands off the children.

For teenagers, the attraction is the power dynamic. Teens have a rather distorted view on what is attractive, and they tend to be naive and easily manipulated. On top of this, almost all teenagers have next to no impulse control, and many will make very very bad decisions (even knowing that the decision is bad) if doing so might result in some form of dopamine hit via sex/adrenaline rush/video games/peer approval/etc. Adults that seek out teenagers for sexual relationships are bad people who chose to be a groomer. There is no genetic component to being a groomer, and they don't deserve pity.

Btw, I can flesh out my claim about the alpaca if you want, but it will have to have a tw for adorable fluffy animals suffering a horrifically slow and painful death.

[–] Adalast@lemmy.world 5 points 1 year ago

Info link: https://pubmed.ncbi.nlm.nih.gov/18686026/

The DSM-V specifies 2 types of pedophilia, Pedophilic (victim age <11) and Hebophilic (victim ages 11-14). What you are describing for the grooming is generally not pedophilia because "children" older than 15 are generally considered post-pubescent and thus anatomically adults. Their frontal lobes still have a LOT of time needed to cook to completion, but they have the impulse control issues for a reason, from an evolutionary standpoint. Yes, in modern society, "adults" who take advantage of the still-developing prefrontal cortex of a post-pubescent adolescent is a shit human being who doesn't deserve to be a member of society, but they are technically not pedophiles, at least not clinically. Legally is a different story, but that is not a pertinent area of discussion right now.

Pedophilic and Hebophilic individuals generally do not ever take their impulses to the realm of reality. Most of them actually end up feeling so much shame and remorse over even having the thoughts that they commit suicide. They definitely deserve pity and treatment, not stigmatization and ostracization.

As to the OP asking about AI art that depicts underage individuals in states of undress or sexual situations, ALL depictions of underage individuals in those contexts are illegal. By the letter of the law, if you draw stick figures on a piece of paper having sex, then label them as children, you have created child pornography. No depiction is legal, no matter the medium. AI-generated, hand drawn, sculpted, watercolors, photos, under the law in (I believe) every state, they are all identical. Personally, I believe that this is asinine and 100% indicates that the purpose of these laws are to adjudicate morality, not "protect the children" as all of the people who push on them claim, but that is just my opinion. Hand-drawn artwork that has no photographic source material and does not depict real people has virtually 0 chance of having caused harm to any children, and AI just knows what the keywords mean in the context of reversing the vaporization of an image. They weren't trained on kiddy porn, the we're trained on pictures of children, and pictures of adults doing their porny thing, so they are able to synthesize the two concepts together.

[–] Ganbat@lemmyonline.com 7 points 1 year ago* (last edited 1 year ago)

Should AI generated CSAM and CP be treated the same as a real person since it promotes the same issues?

That's where things get difficult. An episode of Law & Order: SVU tried to tackle this question a long time ago (but with Photoshopped fake CSAM) and the answer was a resounding "I dunno."

On the one hand, it's disgusting, deplorable, etc. On the other, a fake image means no one was victimized for it.

Does the content further radicalize these people, creating further risk of them victimizing a child, or does it sate their desires, helping to prevent them from victimizing a child? These questions are incredibly difficult to actually answer, and no answer can ever really be definitive, as you can't really predict how any one person might react.

[–] Candelestine@lemmy.world 6 points 1 year ago (2 children)

I don't know many of the answers to these questions, I'm no social scientist or doctor. On the tech side of things, this is all very new and we are still coming to grips with it.

I feel pretty comfortable fielding this one though: There should be no exceptions granted for AI generated pornographic content, and a person's facsimile should have the same protections as actual photos of them. I do not think many people will find this controversial among the general public, as it would pretty clearly serve to protect all of us from having our identities used against our will.

I expect that even our congress should be able to get at least something on the books in this direction. Eventually. Maybe. Or at least the FCC or something.

[–] Delphia@lemmy.world 21 points 1 year ago (9 children)

Let me preface this by saying I DO NOT SUPPORT CSAM!

The only issue I take with AI generated images is that theres no true "age" to the picture and any legislation that would allow people to be jailed or charged would have to be worded very carefully.

"Depicting clearly underage subject matter if it were a person or using prompts to generate someone who clearly appears under aged" simply because someone could be marked for life for typing in "naked elf" and the program spits out something with small boobs and childlike features and not having their HD shredded immediately.

[–] cedarmesa@lemmy.world 16 points 1 year ago* (last edited 1 year ago) (2 children)
[–] Delphia@lemmy.world 14 points 1 year ago (1 children)

Thats exactly my point. Sure the courts may rule in your favor eventually but you just got marched out of work in handcuffs for possession of CSAM, your entire personal and professional circle knows and any explanation you offer is going to sound like total bullshit.

"It was an AI generated image, and it was an elf! She just looked young, but not like illegal young! Guys you have to believe me!"

load more comments (1 replies)
[–] CeruleanRuin@lemmings.world 3 points 1 year ago* (last edited 1 year ago) (1 children)

"Sir, I'm detecting a strong odor of weeb coming from this vehicle."

load more comments (1 replies)
load more comments (8 replies)
load more comments (1 replies)
[–] TheSpermWhale@lemmy.world 3 points 1 year ago

I don’t think the attraction is logical, these people are just mentally ill

[–] dameoutlaw@lemmy.ml 3 points 1 year ago (1 children)

AI CSAM should absolutely be treated as such. The model has been trained on images of real human children. I’m not sure where the issue comes from I would imagine power. Id need to check peer reviewed work from those in the field but I honestly can’t stomach it.

[–] surewhynotlem@lemmy.world 15 points 1 year ago (14 children)

What about an artist just drawing it? Is that ok?

Or no, because the artist has seen children before?

load more comments (14 replies)
load more comments
view more: next ›