this post was submitted on 28 Sep 2023
65 points (86.5% liked)

No Stupid Questions

35966 readers
1291 users here now

No such thing. Ask away!

!nostupidquestions is a community dedicated to being helpful and answering each others' questions on various topics.

The rules for posting and commenting, besides the rules defined here for lemmy.world, are as follows:

Rules (interactive)


Rule 1- All posts must be legitimate questions. All post titles must include a question.

All posts must be legitimate questions, and all post titles must include a question. Questions that are joke or trolling questions, memes, song lyrics as title, etc. are not allowed here. See Rule 6 for all exceptions.



Rule 2- Your question subject cannot be illegal or NSFW material.

Your question subject cannot be illegal or NSFW material. You will be warned first, banned second.



Rule 3- Do not seek mental, medical and professional help here.

Do not seek mental, medical and professional help here. Breaking this rule will not get you or your post removed, but it will put you at risk, and possibly in danger.



Rule 4- No self promotion or upvote-farming of any kind.

That's it.



Rule 5- No baiting or sealioning or promoting an agenda.

Questions which, instead of being of an innocuous nature, are specifically intended (based on reports and in the opinion of our crack moderation team) to bait users into ideological wars on charged political topics will be removed and the authors warned - or banned - depending on severity.



Rule 6- Regarding META posts and joke questions.

Provided it is about the community itself, you may post non-question posts using the [META] tag on your post title.

On fridays, you are allowed to post meme and troll questions, on the condition that it's in text format only, and conforms with our other rules. These posts MUST include the [NSQ Friday] tag in their title.

If you post a serious question on friday and are looking only for legitimate answers, then please include the [Serious] tag on your post. Irrelevant replies will then be removed by moderators.



Rule 7- You can't intentionally annoy, mock, or harass other members.

If you intentionally annoy, mock, harass, or discriminate against any individual member, you will be removed.

Likewise, if you are a member, sympathiser or a resemblant of a movement that is known to largely hate, mock, discriminate against, and/or want to take lives of a group of people, and you were provably vocal about your hate, then you will be banned on sight.



Rule 8- All comments should try to stay relevant to their parent content.



Rule 9- Reposts from other platforms are not allowed.

Let everyone have their own content.



Rule 10- Majority of bots aren't allowed to participate here.



Credits

Our breathtaking icon was bestowed upon us by @Cevilia!

The greatest banner of all time: by @TheOneWithTheHair!

founded 1 year ago
MODERATORS
 

There has been a ton of CSAM and CP arrests in the US lately, especially from cops and teachers, along with at least one female teacher seducing boys as young as 12. I cannot understand the attraction to kids. Even teens. Do these people think they are having a relationship, or it is somehow okay to take away another human beings' innocence? Is it about sex, power, or WTH is it? Should AI generated CSAM and CP be treated the same as a real person since it promotes the same issues? I am a grandfather, and I am worried about how far this will go with the new AI being able to put anyone's face into a Porno movie too.

It seems to me that a whole new set of worldwide guidelines and laws need to be put into effect asap.

How difficult would it be for AI photo apps to filter out words, so someone cannot make anyone naked?

all 40 comments
sorted by: hot top controversial new old
[–] Veraticus@lib.lgbt 58 points 1 year ago* (last edited 1 year ago) (4 children)

I'm only going to answer the first part of your question, not the AI/generated part.

No one really chooses what or who they're attracted to; it kind of just happens to you. For example, you might be watching a TV show and someone gets lightly, comically spanked... and suddenly a light bulb goes off above your head and you think, "whoa, that might actually be kinda fun." People are wired in ways we don't understand to want things we don't even know we want.

To that extent, pedophiles are themselves victims of their own desires; there's no "logic" behind it. It's simply an urge they experience.

Of course that doesn't make succumbing to this urge excusable, and any children who are impacted are of course victims and the pedophiles, predators. But no one is training pedophiles in pedophile camp. It's just humans being human, unfortunately.

[–] DingoBilly@lemmy.world 7 points 1 year ago (2 children)

This is the most accurate answer, and the fact it's all cultural/social is quite important as well.

If you were born a few thousand years ago it may be completely reasonable to sleep with a kid. Hell the kid is probably your slave so you could literally do whatever you want with them.

But just as I don't understand certain fetishes or even just people attracted to the same sex, others won't understand why people would be attracted to kids.

[–] HelixDab2@lemm.ee 2 points 1 year ago (1 children)

If you were born a few thousand years ago it may be completely reasonable to sleep with a kid.

That simply is not correct. While kids have been having sex with kids since, like, forever, looking back to medieval Europe, or even Rome, it was not normal for adults to be having sex with children, especially prepubescent children. Marriage ages for people that weren't nobility (e.g., getting married off to solidify political alliances) were typically early 20s. Children getting pregnant has always been a very dangerous proposition, since it's much more likely to lead to maternal mortality.

Part of the confusion about all of this is that we think people used to live compressed lives, since we see average lifespans in the 40s. But that's not taking into account the ferocious infant mortality rates; if you survived into your late teens, you were probably going to make it to your late 60s/early 70s, even in the middle ages. We think that, since 'average' life spans were 40-odd years, that people must have been marrying very young, but the evidence doesn't really bear that out.

[–] DingoBilly@lemmy.world 0 points 1 year ago

You are incorrect sorry.

While we are on Lemmy, this Reddit AskHistorians post explains how common it was in Ancient Rome. AskHistorians is highly regarded as a good source of information. https://www.reddit.com/r/AskHistorians/comments/feeuye/how_was_pedophilia_considered_in_the_roman_empire/

[–] Firipu@startrek.website -5 points 1 year ago (1 children)

Lol, did you really compare same sex attraction to pedophiles there? "I don't understand both"... Wtf dude.

[–] DingoBilly@lemmy.world 1 points 1 year ago

You do realize that up until a few decades ago in most counties (and even today in some western countries) homosexuality was seen as immoral and illegal?

So yes, it's very comparable in that it's highly socially and culturally defined.

[–] XbSuper@lemmy.world 7 points 1 year ago

This is one of the most sane responses I've ever seen.

I am one of those poor souls who has these urges, but has never, and will never, act on them.

I'm willing to open myself to an AMA for anyone interested.

[–] Hjalamanger@feddit.nu 1 points 1 year ago

I listened to an amazing podcast about this a while ago. It was some science dude helping people not to be attracted to children. If somebody wants to have a listen I can probably find a link, but the podcast was in Swedish

[–] dameoutlaw@lemmy.ml 32 points 1 year ago (1 children)

I see people here attempting to equate it with a natural attraction and or a fetish. We all have internet access and can look it up. It is in the DSM-5 and ICD-10 both pedophilia and pedophilic disorder. Especially, in this more advanced age of medicine, science and society. If it was natural I believe the corrections would’ve been made and or strongly advocated for, it’s not. What is advocated is using terminology correctly, encouraging those that experience this to feel comfortable to tell their truth and seek help. I believe some of you guys comments are very dangerous and some of the upvotes and downvotes are concerning and makes it difficult to distinguish if you are in support of protecting children. The point is please don’t just blanket label it and compare it to things that aren’t harmful, illegal and consensual.

[–] HelixDab2@lemm.ee 2 points 1 year ago

If it was natural I believe the corrections would’ve been made and or strongly advocated for, it’s not.

You mean corrections to the DSM? There are lots of things that are natural and have organic causes that are still in the DSM, and will remain in the DSM, because they're problems for the individual and society. Homosexuality was removed because it was recognized that acting on that sexual orientation, with another adult that had the same sexual orientation, was not a problem for society or the individuals. You can't do the same with child sexual abuse; children can't consent, legally, morally, or ethically.

[–] HelixDab2@lemm.ee 27 points 1 year ago (2 children)

There are multiple parts to your question. I'm going to try to break it down.

First, there's a difference between a pedophile and a child molester. Pedophilia is a sexual attraction to children, but it does not, by itself, require the person to take action. A child molester is a person that sexually assaults children. It's the difference between being heterosexual, and being a rapist; you can be straight and still be entirely celibate.

Child molesters may not be sexually attracted to children at all; some might be, but people that commit rape aren't usually doing it solely for sexual gratification, although sex is definitely part of it.

We don't know how common pedophilia is because of how heavily stigmatized it is.

You don't understand how a person could be sexually attracted to children; the simplest way to explain it is to ask if you can understand how a man can be sexually attracted to another man. IIRC, most research indicates that pedophilia probably is a sexual orientation, much like being straight, gay, or bisexual is (except that there is no moral or ethical way for a pedophile to have a sexual or romantic relationship with a child; that is always both predatory and criminal). Do pedophile child molesters believe that they're having a relationship? Some of them, yes. They're able to delude themselves into believing that the child wants the attention and sex (really sexual assault), when they're--probably--the one that has groomed the child in the first place.

I cannot understand the attraction to kids. Even teens.

I can. When I was a child, I was sexually attracted to my peers. 14yo kids are having sex with each other, so clearly they're attracted to each other. As an adult, I can see women in their 20s as being sexually attractive, while still having zero interest in them (y'all seem really young, and not in a good way, if y'know what I mean). Sexual maturity isn't a magical thing that happens when you hit 18 (or whatever the age of consent is where you live); it's a sliding scale.

Should AI generated CSAM and CP be treated the same as a real person since it promotes the same issues?

I don't think that you can make a person into a pedophile, any more than you can make a person gay. A person either is, or isn't, a pedophile, and CG CSAM isn't going to change that. So the question is, does CG CSAM make it more likely that a pedophile will end up sexually abusing a child? My intuition says that it will not, in the same way that the proliferation of pornography has not made sexual assault of adults more common. (Some research indicates that the availability of pornography has decreased rates of sexual assault.) Child pornography is illegal--in part--because it cannot be produced without causing real harm to children. CG CSAM doesn't cause real harm to any person though; unless there's evidence that it increases the rates of child sexual abuse, I don't think that the squick factor is a reasonable basis for banning it. OTOH, adult pornography has generally led to a relaxation of sexual mores and norms--which I believe is a generally positive thing--and it's possible that CG CSAM would normalize child sexual abuse sufficiently that libertarians would be able to severely weaken age of consent and statutory rape laws. I don't really know, TBH; I'd want to see more research rather than reflexively banning it.

[–] MystikIncarnate@lemmy.ca 3 points 1 year ago (1 children)

Articulate and concise. I like it.

I don't really have anything more to add, except that we as a society need to make up our collective minds on cg and AI csam/CP soon.

I definitely think that cg/AI content is less bad (still bad, but less so) because there's no harm being done to real children from it; but those AI image engines needed to be trained on some form of content to be able to generate the images that they do, so I'm not sure how the training images factor in, or what would even be used for training that AI... I know rather little about how AI is trained at the moment, so I'm not sure if it can be done without source csam material or not... IMO, that factors into the morality of the output.

I definitely agree that sexual preference toward minors (aka paedophilia) is just that, a sexual preference; and that, in and of itself does not make someone a sex offender in the same way that being heterosexual doesn't make you a rapist or other form of sexual "deviant" (or however you want to say that. It's interesting to me to think that paedophiles may have a semi-legal way of getting porn for themselves (which causes no harm to children). I feel a bit bad for paedophiles in that they're basically forced to have relationships with persons that they don't find very sexually attractive, else they break the law. Not bad enough that I think that laws should change our anything, it's just a crap situation. It would be like having a preference towards men, as a man, in a world of heteros. The men are there and you're interested in them, but none of them are interested in you. Almost always that's not the case, there's other homosexual men that exist, no matter how rarely... in the case with pedos, there are exactly zero underage people who they can interact with at all sexually. I still don't think that should change, but at least with the internet, a gay man can go and find porn that interests them. With pedos is literally a crime to even look at, possess, or make any porn that appeals to them.

I can sympathize with the impossibility of their situation, that's all. For the record, I'm just done cis male with no interest in anyone too young to date. I can recognise their attractive qualities without being attracted to them (speaking mostly about those that have reached their sexual maturity here, who are still not 18 or whatever)... I can understand it, I'm not so hateful to want anyone who feels attraction to young people to die or anything, but young people don't have the experience to understand the situation they're getting into, when they're being mislead or gas lit, etc (though to be fair, a lot of "mature people seem to not know either, but that's another discussion)... Fact is, they're shit out of luck.

I'm sure many are forced into celibacy just to be lawful. I don't think any grown adult wants to be forced to be celibate; so I can understand the plight. AI/cg porn, tailored to that specific preference may give pedos an outlet that they can utilize to temper their urges and keep them on the right side of the law here. Of course it won't solve the problem entirely, the same way that rapists are still a thing, but it may severely reduce illegal activity and harm to children.

But I agree, it's a slippery slope (so to speak) because it can easily evolve into lowering the age of consent, and bringing back child marriages and such. Which IMO, isn't a desired outcome. I also don't think that content should intermingle with either social networks or existing porn sites, since it's so specific, it should be relegated to specific sites and not left flapping around the internet. It's also a vast minority of people that are afflicted, so segregation may be a minimum measure to keep things somewhat clean. I know I don't want AI generated CP content mixed in with my usual porn browsing... I'm sure there's plenty of people in the same boat, so IMO that's a minimum. But I'm only one voice in the society, so I don't make the decision; I'm interested to see what decision is finally made and implemented, whenever we get there.

As a disclaimer: I'm not attracted to underage people. I'm also not a doctor or scientist, or psychologist or anything else. I'm not in favor of anything here, besides society making a decision, and I'm just positing that it could be beneficial to society as a whole. I welcome other opinions, except those by people whom are heavily religious. Good day.

[–] HelixDab2@lemm.ee 2 points 1 year ago (1 children)

I feel a bit bad for paedophiles in that they’re basically forced to have relationships with persons that they don’t find very sexually attractive, else they break the law.

I've seen a paper--which I unfortunately did not bookmark--that seemed to indicate that most pedophiles were not exclusively pedophiles; many are able to have romantic and sexual relationships with age-appropriate partners. They also tend to have a distinct gender preference for minors (e.g., a person that is a heterosexual and a pedophile will prefer minors that are in-line with their sexual orientation). The ones that are 'pure' pedophiles--not sexually attracted to any adults at all--do not seem to have a gender preference, which kinda makes sense when you consider secondary sex characteristics as markers of physical maturity, e.g., young boys and girls look physically very similar aside from the genitals themselves. Again - I don't have the reference on this saved, so I might be misremembering, or misrepresenting it, but this is what I recall.

there are exactly zero underage people who they can interact with at all sexually.

There's a genetic disorder--I believe exclusively in women--where they don't 'grow up'; they don't get very tall, they're largely lacking in secondary sexual characteristics, and I believe that they're infertile. I ran into a woman like that--who was with her partner--at a fetish event. It really gave me whiplash, because at first, second, and third glances she looked like she was 12, at an event that had explicit sexual activity in the open. It took a closer look at her face to realize that she was in her 30s. God bless her, she found someone that was attracted to her, and into the same shit she was into. So, y'know, there's that.

[–] MystikIncarnate@lemmy.ca 3 points 1 year ago (1 children)

I've heard of that genetic condition. It's fascinating, and as far as I know, extremely rare.

I know that at least one has spoken publicly about her experience, and they touched on dating and the implication was that most of the people that are interested, are paedophiles, and that didn't sit well for her, and I expect that wouldn't sit well for most people, especially those with that condition.

Fascinating information all around. I don't have a doubt that is accurate.

[–] HelixDab2@lemm.ee 0 points 1 year ago (1 children)

She's got a catch-22 there; if she doesn't want to date anyone that's attracted to her because they're likely a pedophile, then she's not going to ever be able to have any romantic relationships (assuming that she wants them). I guess if that were me, I'd rather date a person that was sexually and romantically attracted to me--despite knowing that they were also sexually attracted to minors--than live life completely alone.

[–] MystikIncarnate@lemmy.ca 2 points 1 year ago

That was the take away. She was rather upset about it, which is apparently good for ratings but tragic overall.

I suppose it depends on what she really wants in life, which I won't presume to know. I wish her the best, that's not a fun condition to deal with.

[–] Chailles@lemmy.world 2 points 1 year ago

I don’t think that you can make a person into a pedophile, any more than you can make a person gay.

If you really think about it, we've seen arguments like that before. That pornography creates rapists. That violent video games creates murderers. And that's just strictly on the consumption of media.

[–] Astroturfed@lemmy.world 22 points 1 year ago

Don't try to understand it. You aren't going to get a good answer. It's a horrible mental illness level of sexual preference.

Anything can be sexualized with enough impulse and experiences. Everyone's got some weird dark fetish shit. Some of it's illegal in practice. Normal people bury that shit or only discuss it in therapy. While talking it out so they can hopefully never think about it again.

I'm sure there's different answers to this just like "why are there serial killers?". Just be glad it confuses you.

[–] scarabic@lemmy.world 13 points 1 year ago

I imagine there’s some kind of vampiric quality to it. Kids are full of youth and innocence: things we are all constantly losing to time. Especially if someone’s own childhood was robbed from them, I think they will carry around a void they desperately want to fill but never can, because of course abusing a child doesn’t bring these things back to you. Still, many child abuse victims go on to abuse other children later in life, and this may be their drive: to seek the thing they lost. It’s beyond sad. Abusing children is straight up disgusting and terrible, but the convoluted desperation that causes people to do it is truly horrifying in a stranger-than-any-fiction kind of way.

[–] Apepollo11@lemmy.world 12 points 1 year ago* (last edited 1 year ago) (1 children)

I'm only going to tackle the tech side of this...

How difficult would it be for AI photo apps to filter out words, so someone cannot make anyone naked?

Easy. The most popular apps all filter for keywords, and I know that at least some then check the output against certain blacklisted criteria to make sure it hasn't let something slip through.

But...

Anyone can host their own version and disable these features, allowing them to generate whatever they want, in the exactly same way that anyone can write their own story containing whatever they want. All you need is the determination to do it, and some modicum of ability.

People have been been creating dodgy doctored photos long before computers. When Photoshop came out, it became easier, and with AI it's easier still. The current laws about creating and distributing indecent images still apply to these new images though.

[–] Adalast@lemmy.world 2 points 1 year ago (1 children)

Technically the diffusers all have the ability to filter any material from the actual outputs using a secondary CLIP analysis to see if it kicks out any keywords which indicate that a topic is in the image. From what I have seen, most AI generation sites use this method as it is more reliable for picking up on naughty outputs than prompt analysis. AI's are horny, I play with it a lot. All you have to do is generate a woman on the beach and about 20% of them will be at least topless. Now, "woman on the beach" should not he flagged as inappropriate, and I don't believe the outputs should either because our demonization of the female nipple is an asinine holdover from a bunch of religious outcasts from Europe who were chased our for being TOO restrictive and prudish, but alas, we are stuck with it.

[–] skullgiver@popplesburger.hilciferous.nl 1 points 1 year ago* (last edited 1 year ago) (1 children)

[This comment has been deleted by an automated system]

[–] Adalast@lemmy.world 1 points 1 year ago

You are correct, CLIP can misinterpret things, which is where human intelligence comes in. Having CLIP process the probabilities for the terminology that you describe what you are looking for then utilizing a bit of heuristics can go a long way. You don't need to train it to recognize a nude child because it has been trained to recognize a child, and it has been trained to recognize nudity, so if an image scores high in "nude" and "child" just throw it out. Granted, it might be a picture of a woman breastfeeding while a toddler looks on, which is inherently not child pornography, but unless that is the specific image that is being prompted for, it is not that big of a deal to just toss it. We understand the conceptual linking so we can set the threshold parameters and adjust as needed.

As for the companies, it is a tough world surrounding it. The argument of a company that produced a piece of software being culpable for the misuse of said software is a very tenuous one. There have been attempts to make gun manufacturers liable for gun deaths (especially handguns since they really only have the purpose of killing humans). This one I can see, as the firearm killing a person is not a "misuse", indeed, it is the express purpose for it's creation. But what this would be would be more akin to wanting to hold Adobe liable for the child pornography that is edited in Lightroom, or Dropbox liable for someone using Dropbox API to set up a private distribution network for illicit materials. In reality, as long as the company did not design a product with the illegal activity expressly in mind, then they really shouldn't be culpable for how people use it once it is in the wild.

I do feel like more needs to be done to make public the training data for public inspection, as well as forensic methods for interrogating the end products to figure out if they are lying and hiding materials that were used for training. That is just a general issue though that covers many of the ethical and legal issues surrounding AI training.

[–] ThisIsAManWhoKnowsHowToGling@lemmy.dbzer0.com 11 points 1 year ago (1 children)

There are two parts to this problem.

For kids who haven't hit puberty, there is a diagnosable pedophilia disorder. This is mostly genetics. (I'm pretty sure I've met an alpaca that was a pedophile once.) The molester's brain is wired wrong. Nothing to do about that. IMHO, they deserve pity as long as they keep their hands off the children.

For teenagers, the attraction is the power dynamic. Teens have a rather distorted view on what is attractive, and they tend to be naive and easily manipulated. On top of this, almost all teenagers have next to no impulse control, and many will make very very bad decisions (even knowing that the decision is bad) if doing so might result in some form of dopamine hit via sex/adrenaline rush/video games/peer approval/etc. Adults that seek out teenagers for sexual relationships are bad people who chose to be a groomer. There is no genetic component to being a groomer, and they don't deserve pity.

Btw, I can flesh out my claim about the alpaca if you want, but it will have to have a tw for adorable fluffy animals suffering a horrifically slow and painful death.

[–] Adalast@lemmy.world 5 points 1 year ago

Info link: https://pubmed.ncbi.nlm.nih.gov/18686026/

The DSM-V specifies 2 types of pedophilia, Pedophilic (victim age <11) and Hebophilic (victim ages 11-14). What you are describing for the grooming is generally not pedophilia because "children" older than 15 are generally considered post-pubescent and thus anatomically adults. Their frontal lobes still have a LOT of time needed to cook to completion, but they have the impulse control issues for a reason, from an evolutionary standpoint. Yes, in modern society, "adults" who take advantage of the still-developing prefrontal cortex of a post-pubescent adolescent is a shit human being who doesn't deserve to be a member of society, but they are technically not pedophiles, at least not clinically. Legally is a different story, but that is not a pertinent area of discussion right now.

Pedophilic and Hebophilic individuals generally do not ever take their impulses to the realm of reality. Most of them actually end up feeling so much shame and remorse over even having the thoughts that they commit suicide. They definitely deserve pity and treatment, not stigmatization and ostracization.

As to the OP asking about AI art that depicts underage individuals in states of undress or sexual situations, ALL depictions of underage individuals in those contexts are illegal. By the letter of the law, if you draw stick figures on a piece of paper having sex, then label them as children, you have created child pornography. No depiction is legal, no matter the medium. AI-generated, hand drawn, sculpted, watercolors, photos, under the law in (I believe) every state, they are all identical. Personally, I believe that this is asinine and 100% indicates that the purpose of these laws are to adjudicate morality, not "protect the children" as all of the people who push on them claim, but that is just my opinion. Hand-drawn artwork that has no photographic source material and does not depict real people has virtually 0 chance of having caused harm to any children, and AI just knows what the keywords mean in the context of reversing the vaporization of an image. They weren't trained on kiddy porn, the we're trained on pictures of children, and pictures of adults doing their porny thing, so they are able to synthesize the two concepts together.

[–] Ganbat@lemmyonline.com 8 points 1 year ago* (last edited 1 year ago)

Should AI generated CSAM and CP be treated the same as a real person since it promotes the same issues?

That's where things get difficult. An episode of Law & Order: SVU tried to tackle this question a long time ago (but with Photoshopped fake CSAM) and the answer was a resounding "I dunno."

On the one hand, it's disgusting, deplorable, etc. On the other, a fake image means no one was victimized for it.

Does the content further radicalize these people, creating further risk of them victimizing a child, or does it sate their desires, helping to prevent them from victimizing a child? These questions are incredibly difficult to actually answer, and no answer can ever really be definitive, as you can't really predict how any one person might react.

[–] Izzgo@kbin.social 5 points 1 year ago (1 children)

To me there is a clear difference between children, and teens say 16+. It is both morally wrong and unnatural to be attracted to prepubescent children, and this is pedophilia. But basically, by definition puberty makes people become sexually attractive, and it's natural for adults to be attracted. Still morally wrong to act on those attractions unless you're in about the same stage of puberty or early adulthood. That's when we rely on a strong moral code and laws in society to protect youngsters who have recently gone through puberty. And hopefully even after the laws no longer apply, we have enough societal pressure to strongly discourage wide age gaps between sexual partners.

Pedophilic disorder is characterized by recurring, intense sexually arousing fantasies, urges, or behavior involving children (usually 13 years old or younger).

[–] dameoutlaw@lemmy.ml 3 points 1 year ago

There is a term for that Ephebophilia generally covers ages 15 to 18

[–] cooopsspace@infosec.pub 2 points 1 year ago* (last edited 1 year ago)

At the end of the day, art is just pixels on a flat surface. Determining whether a depicted individual is under age where it's not obvious sets a dangerous precedent. Is the picture 17 or 18? Who knows.

But the problem is that people have been sexualising people like Emma Watson since she first appeared on screen. That's not okay and rather than sending AI art underground I think society needs to change to normalise education about sex, reproduction and genitalia and address the social issues to treat pedophilia like the disease that it is.

Meanwhile pedophile names are being written about publicly, risking mob violence and further isolation. Not to mention in the US theres a lot of negative attention being put on women's reproduction, childrens sex ed and genitalia and a push to make the whole lot illegal and taboo. Not to mention people teaching their kids pet names for their parts, "uncle Ben touched my heehaw" sounds a lot different to "uncle Ben touched my penis".

Society is a problem, the US particularly is going in the wrong direction on many aspects of sex education.

[–] Roflmasterbigpimp@lemmy.world 0 points 1 year ago

I'm not entirely sure about this but as far as I know is there also a genetic Component to it.

[–] BastianAI@lemmy.world 0 points 1 year ago (2 children)

AI generated CSAM and CP should be treated the same, because you can't generate anything without first having trained the model on the kind of content you want to generate.

[–] fubo@lemmy.world 23 points 1 year ago* (last edited 1 year ago) (1 children)

There are no green cats in the training data, but the AI can interpolate between a green parrot and a gray cat.

[–] WHYAREWEALLCAPS@lemmy.world 6 points 1 year ago

Precisely. It's a mashup of what it has been trained on. Train it on celebrities in bikinis and pornstars getting railed in bikinis, smash the two together and et voila, suddenly every actress has a porn tape of getting railed in a bikini.

[–] gullible@kbin.social -1 points 1 year ago

Pedophilia is bad.

[–] dameoutlaw@lemmy.ml -1 points 1 year ago

AI CSAM should absolutely be treated as such. The model has been trained on images of real human children. I’m not sure where the issue comes from I would imagine power. Id need to check peer reviewed work from those in the field but I honestly can’t stomach it.