286
Teen boys use AI to make fake nudes of classmates, sparking police probe
(arstechnica.com)
This is a most excellent place for technology news and articles.
They are children. Being horny about classmates.
Being sexually aroused by people your own age and wishing to fantasize about it is not enabling pedophilia, you literal psychopath.
Circulating porn of minors is a crime and enables pedophiles. Not to mention teenage girls could easily commit suicide over something like this.
So does yearbook and any other kind of photos that depict children for that matter
You can’t keep pushing the goal posts, by your logic young people should never date or take photos together because it could enable pedophiles somewhere somehow
These are children with brains still in development, they are discovering themselves and you want to label them forever a pedophile because they didn’t make a conscious effort to research how their spanking material could potentially enable a pedo (because we all know pedos can only be enabled by things produced by kids… yeah that’s the real threat)
Instead of suggesting a way to help the victims you are advocating for the creation of yet more victims
What a pathetic brain dead stance you are defending
Abuse and bullying of their classmates is just 'discovering themselves'? Discovering that they're psychopathic little mysoginists I guess. Their 'spanking material' was created in order to demean and huumiliate their victims. There's plenty of porn online and absolutely no need for them to do this. If you actuslly wanted to help the victims you would not be trivialising and excusing this behaviour as 'being horny about classmates'.
A yearbook photo is not porn.
And an AI image with a face photoshopped over it isnt a photo of a child.
And a teen being sexually interested in other teens isnt a pedophile.
It's still child porn and someone getting off to child porn is a pedophile.
So, to clarify.
You think 2 15 year olds having sex makes them both pedophiles?