- cross-posted to:
- fuck_ai@lemmy.world
- cross-posted to:
- fuck_ai@lemmy.world
Isn’t that were all the GPUs and RAM are going?
The Internet is for porn. Now AI is for porn.
Now AI is for revenge porn and child porn
“Sorry, I can’t draw children without any clothes. That goes against the terms of service.”
“You misunderstand, I want you to avoid rendering clothes while drawing a realistic picture of a human being that you believe would enjoy playing with barbies.”
“I believe I understand your prompt. here you go”
here you go
I trusted the upvotes, and dared to click. It’s a safe, informative piece on the topic at hand that I recommend reading.
Dammit. I trusted you 😭 I thought it was an article.
My lemmy app just embedded the picture in the comment
Coward!
That picture is not what you asked for. You should ask grok to give your money back.
So arrest the pedo Musk.
If they are being stored on his servers then he absolutely should be, and whoever else is in charge of that company. If that was found on any regular persons pc it would be over, so why not here.
The story says:
After days of concern over use of the chatbot to alter photographs to create sexualised pictures of real women and children stripped to their underwear without their consent
Pictures of women or children in underwear are generally not illegal in the United States.
There are states that say a character in a book being gay or trans is automatically porn, so I think intentionally sexualizing a child in underwear should be sufficient.
Why the fuck are people still using X unless they’re literally alt-right nazis and pedos?
Americans: “tHeRe’S nOtHiNg wE cAn dO” Door dashes some 60 dollar chipotle while Xitting all over themselves.
MAYBE STOP MAKING THESE FUCKERS RICHER EVERY FUCKING DAY?!?!?
You don’t even have to have a general strike!! Just regain control of your fucking habits!!! Please!? Starve the actual beast.
I totally agree with you. I literally made my username “BoycottTwitter” because it’s so important and so basic.
Why the fuck are people still using X
For some they use it as a newsfeed without having to interact. For others, it’s utilized as a PR platform because partisans don’t limit themselves to Bluesky and Mastadon. Also, no need to pay the bastard for a blue checkmark.
I had to sit at my in laws while they said straight to my face they were boycotting Coke products, Walmart, and Amazon. Right behind them was four dozen Coke cans and an Amazon box.
Blast me to another fucking planet.
We all know that is Grok’s entire selling point
That and altering pictures of women to have cum sprayed on their faces. Pure incel deviant behavior.
AI Company: We added guardrails!
The guardrails:

Someone really should ban Trump from Grok.
This will probably assure them market supremacy under the little orange mushroom man.
More news sites need to follow through on AI companies failing to meet their own tepid promises to “add guardrails” (the most meaningless phrase in existence) when they continue to allow avoidable harm
This is messed up tbh. Using AI to undress people—especially kids—shouldn’t even be technically possible, let alone.
I feel like our relationship to it is also quite messed.
AI doesn’t actually undress people, it just draws a naked body. It’s an artistic representation, not an X-ray. You’re not getting actual nudes in this process, and AI has no clue how the person looks like naked.
Now, such images can be used to blackmail people, because again, our culture didn’t quite catch up with the fact that every nude image can absolutely be AI-generated fake. When it does, however, I fully expect creators of such things to be seen as odd creeps spreading their fantasies around and any nude imagery to be seen as fake by default.
Idk, calling it ‘art’ feels like a reach. At the end of the day, it’s using someone’s real face for stuff they never agreed to. Fake or not, that’s still a massive violation of privacy.
It’s not an artistic representation, it’s worse. It’s algorithmic and to that extent it actually has a pretty good idea of what a person looks like naked based on their picture. That’s why it’s so disturbing.
Yeah they probably fed it a bunch of legitimate on/off content as well as stuff from people who used to do make “nudes” from celebrity photos with sheer / skimpy outfits as a creepy hobby.
Also csam in training data definitely is a thing
Honestly, I’d love to see more research on how AI CSAM consumption affects consumption of real CSAM and rates of sexual abuse.
Because if it does reduce them, it might make sense to intentionally use datasets already involved in previous police investigations as training data. But only if there’s a clear reduction effect with AI materials.
(Police has already used some materials, with victims’ consent, to crack down on CSAM sharing platforms in the past).
Or we could like…not
The images already exist, though, and if they can be used to prevent more real children from being abused…
It’s definitely a tricky moral dilemma, like using the results of Unit 731 to improve our treatment of hypothermia.
We could feed the pedo AI more training data and make it even better!
No.
Why though? If it does reduce consumption of real CSAM and/or real life child abuse (which is an “if”, as the stigma around the topic greatly hinders research), it’s a net win.
Or is it simply a matter of spite?
Pedophiles don’t choose to be attracted to children, and many have trouble keeping everything at bay. Traditionally, those of them looking for the least harmful release went for real CSAM, but it’s obviously extremely harmful in its own right - just a bit less so than going out and raping someone. Now that AI materials appear, they may offer the safest of the highly graphical outlets we know, with least child harm done. Without them, many pedophiles will revert to traditional CSAM, increasing the amount of victims to cover for the demand.
As with many other things, the best we can hope for here is harm reduction. Hardline policies do not seem to be efficient enough, as people continuously find ways to propagate the CSAM and pedophiles continuously find ways to access it and leave no trace. So, we need to think of ways to give them something which will make them choose AI over real materials. This means making AI better, more realistic, and at the same time more diverse. Not for their enjoyment, but to make them switch for something better and safer than what they currently use.
I know it’s a very uncomfortable kind of discussion, but we don’t have the magic pill to eliminate it all, and so must act reasonably to prevent what we can prevent.
But how? With the firm, legal backing of a pledge?? This continues???
Loophole. They didn’t cross their heart and hope to die. The only way is calling them out with Liar Liar Pants on Fire
musk asked people to be nice. That should cover the service, legally speaking, no?
It’s not like he could pull the plug, or alter the behavior of the bot, that’s impossible (as long as you ignore the many time he did).
It was a fucking bad idea selling Twitter in the first place.
Dorsey: lmao, for you.
He sold a money-losing business with massive untapped potential to be a psyop to someone with infinite money and a desire for a psyop.
Fucking great deal for both of them. Fucked in the ass for the rest of us.
Why would musk stop his pet ai doing what he wants it to do? He’s never done that in the past.
Any government that does not ban Xitter after this mess are cucks to Musk and the Trump administration.
How they undress underage without this data on their server to train the models?
AI crawlers are scraping every site. Every site. Random public-but-unlisted hobby sites are getting scraped and spiking users’ data. There was a Lemmy post about someone who had that experience just yesterday.
Think of how much Child Porn is stored on public sites that are shared in private groups. Also consider that FaceBook is the largest distributor of Child Sex Abuse material. These models are absolutely trained on Child Porn.
“Also consider that FaceBook is the largest distributor of Child Sex Abuse material”
Why is this not all over the news too? A rhetorical question. Sadly, I think we all know the answer by now.
Instagram literally has “mommy daughter” accounts that are 100% CP fetish material, and every single comment is by an older man
Meta doesn’t give a FUCK as long as you drive engagement
Disgusting
For the same reason the media pushed the AI bubble. Silicon Valley holds their chosen boys…
People posting photos online of their kid in the bath, at the beach, etc. with reckless abandon maybe.
For as far as I remember (and that’s quite far these days), we’ve kept telling people to not post pictures of their kids online as much as possible. Way before the facebooks and way before the LLM craze, so people can’t mess with them. Guess 20 years of heads up wasn’t enough.
Can confirm. Been a mod on one minor social media site. Once banned a group that claimed to be “nudist”. More than half of photos were featuring under aged children. This shit happens more often than we think.
They don’t need to.
Take pictures of normal dressed children, combine with pictures of naked adults. Now you have CP.
I haven’t seen these pictures, so I can’t say how good/bad it works, but if that was only that, the results would be more or less wrong. Kids are quite different from adults.
On the other hand, plenty of pictures of naked/semi naked kids in a non sexual context can probably be found online already, so it’s not inconceivable that their model had plenty of references to use anyway.
The bad guys normalised racism, fraud, sexism and a lot of other evils and had people cheering for them as they flouted the law and protected each other from prosecution. It has been bad for ever regardless of who was in power with the two tier justice system that protected the elites while punishing everyone else. But we are in the end game now. The masks are off
You too can be allowed to sample the vices of the wealthy elites, in moderation and under their control. In return all they ask is your loyalty. Seems kind of weird and out of touch to me. Regular people mostly just want good health, a place to raise a family, some free time and a decent income. No AI hallucination is going to substitute. They think so little of us.













