this post was submitted on 19 Sep 2023
622 points (98.0% liked)

Europe

8484 readers
1 users here now

News/Interesting Stories/Beautiful Pictures from Europe 🇪🇺

(Current banner: Thunder mountain, Germany, 🇩🇪 ) Feel free to post submissions for banner pictures

Rules

(This list is obviously incomplete, but it will get expanded when necessary)

  1. Be nice to each other (e.g. No direct insults against each other);
  2. No racism, antisemitism, dehumanisation of minorities or glorification of National Socialism allowed;
  3. No posts linking to mis-information funded by foreign states or billionaires.

Also check out !yurop@lemm.ee

founded 1 year ago
MODERATORS
 

Police investigation remains open. The photo of one of the minors included a fly; that is the logo of Clothoff, the application that is presumably being used to create the images, which promotes its services with the slogan: “Undress anybody with our free service!”

you are viewing a single comment's thread
view the rest of the comments
[–] blargerer@kbin.social 10 points 1 year ago (4 children)

Obviously this is creepy, but the technology is out there, one of those can't put the genie back in the bottle techs. You can and should look at the people generating the images as creeps, but ultimately we as a society need to learn to not put as much veracity or identity in images now.

With that said where the fuck did this model get its training data for 14 year olds. That sounds like a more serious issue.

[–] barsoap@lemm.ee 15 points 1 year ago (2 children)

With that said where the fuck did this model get its training data for 14 year olds.

Nowhere, at least for any model you could get your hands at in public places like civitai. Or, well, it's not like they can tell whether someone trained on those kinds of pictures but they're rightly nuking any underage/loli example images, as well as anyone who posts them, from orbit.

Generally speaking models can be very good at mixing concepts they have an understanding of, say a giraffe with zebra stripes, but that doesn't mean that you can just combine anything -- if you try to generate a nude human with zebra fur you're bound to get body paint, random skimpy zebra-striped clothing, or at most a fursuit, not convincing fur, unless you use a model trained by furries but at that point you'll have trouble generating faces without muzzles: The AI just doesn't know how actual zebrakin look like so it's either copping out or making stuff up.

I've never tried nor am I remotely attracted to that age range but I wouldn't be surprised if a paedophile would complain "these aren't kids they're scaled-down adults". Things like the difference between budding and small breasts, ask a biologist I haven't seen 14yold breasts in over two decades.

On another note though I'd much rather have paedophiles jack off to generated images than doing anything involving actual children, including creeping around. Lesser of two evils and all that. Therapy, of course, is preferable to both.

[–] ParsnipWitch@feddit.de 1 points 1 year ago (2 children)

How do you know that these people replace harassment with these pictures? And not just do both, or even increase their fetishes?

What about the girls who's pictures were used as material for these generated images?

[–] barsoap@lemm.ee 0 points 1 year ago

How do you know that these people replace harassment with these pictures? And not just do both, or even increase their fetishes?

AFAIK psychologists simply don't know, and it might be a case by case thing.

What about the girls who’s pictures were used as material for these generated images?

As I explained, it might not be necessary to have any underage material in the training data.

Generally speaking I didn't come here to have a deep discussion about a very difficult moral and legal issue, I'll leave that up to the specialists. I wanted to say something about AI and somehow all answers I get are about the last tacked-on paragraph making a quick statement about me preferring keeping paedophiles away from kids.

[–] krellor@kbin.social 5 points 1 year ago (1 children)

Right, the technology is out there so we as a society need to establish norms, customs, and yes, laws governing its use.

I'm pretty firmly on the side of there being legal consequences for taking pictures of real minors, running them through a service to create nude replicas, and then circulating those pictures. That is wrong on so many levels and could constitute any number of crimes without the AI component including, such as harassment. I mean, intentionally using someone's likeness to circulate embarrassing materials already had legal consequences. This is just a whole other level of ick on top.

[–] taladar@feddit.de 3 points 1 year ago (2 children)

Personally I don't see a difference between using an AI service or plain old Photoshop to create a fake nude picture of someone. Both should be punished in the same way and if law makers haven't caught up with the Photoshop version after 30 years they likely won't handle the AI version in this century either.

[–] krellor@kbin.social 3 points 1 year ago

I would agree, though I wonder about the service mentioned that is dedicated to the process. My comment was in response to someone who seemed to think circulating fake nudes wasn't a problem, regardless of how they were generated.

[–] ParsnipWitch@feddit.de 1 points 1 year ago

It's not enough to just look at the people who did this as creeps.