this post was submitted on 21 May 2024
78 points (97.6% liked)

Technology

1345 readers
142 users here now

Which posts fit here?

Anything that is at least tangentially connected to the technology, social media platforms, informational technologies and tech policy.


Rules

1. English onlyTitle and associated content has to be in English.
2. Use original linkPost URL should be the original link to the article (even if paywalled) and archived copies left in the body. It allows avoiding duplicate posts when cross-posting.
3. Respectful communicationAll communication has to be respectful of differing opinions, viewpoints, and experiences.
4. InclusivityEveryone is welcome here regardless of age, body size, visible or invisible disability, ethnicity, sex characteristics, gender identity and expression, education, socio-economic status, nationality, personal appearance, race, caste, color, religion, or sexual identity and orientation.
5. Ad hominem attacksAny kind of personal attacks are expressly forbidden. If you can't argue your position without attacking a person's character, you already lost the argument.
6. Off-topic tangentsStay on topic. Keep it relevant.
7. Instance rules may applyIf something is not covered by community rules, but are against lemmy.zip instance rules, they will be enforced.


Companion communities

!globalnews@lemmy.zip
!interestingshare@lemmy.zip


Icon attribution | Banner attribution

founded 11 months ago
MODERATORS
 

cross-posted from: https://lemmy.zip/post/15863526

Steven Anderegg allegedly used the Stable Diffusion AI model to generate photos; if convicted, he could face up to 70 years in prison

top 50 comments
sorted by: hot top controversial new old
[–] Daxtron2@startrek.website 33 points 5 months ago (39 children)

How are they abuse images if no abuse took place to create them?

[–] PoliticalAgitator@lemmy.world 3 points 5 months ago* (last edited 5 months ago) (2 children)

Because they are images of children being graphically raped, a form of abuse. Is an AI generated picture of a tree not a picture of a tree?

[–] Daxtron2@startrek.website 4 points 5 months ago* (last edited 5 months ago) (14 children)

No it isn't, not anymore than a drawing of a car is a real car, or drawings of money are real money.

[–] PoliticalAgitator@lemmy.world 4 points 5 months ago (13 children)

Material showing a child being sexually abused is child sexual abuse material.

load more comments (13 replies)
load more comments (13 replies)
[–] Leg@lemmy.world 3 points 5 months ago (4 children)

It's a picture of a hallucination of a tree. Distinguishing real from unreal ought to be taken more seriously given the direction technology is moving.

load more comments (4 replies)
[–] sxt@lemmy.world 3 points 5 months ago (4 children)

If the model was trained on csam then it is dependent on abuse

[–] Darrell_Winfield@lemmy.world 25 points 5 months ago (1 children)

That's a heck of a slippery slope I just fell down.

If responses generated from AI can be held criminally liable for their training data's crimes, we can all be held liable for all text responses from GPT, since it's being trained on reddit data and likely has access to multiple instances of brigading, swatting, man hunts, etc.

[–] laughterlaughter@lemmy.world 2 points 5 months ago

You just summarized the ongoing ethical concerns experts and common folk alike have been talking about in the past few years.

[–] Daxtron2@startrek.website 19 points 5 months ago

As I said in my other comment, the model does not have to be trained on CSAM to create images like this.

load more comments (2 replies)
[–] mindbleach@sh.itjust.works 2 points 5 months ago (1 children)

All the lemmy.world commenters came out to insist "that painting is a pipe, though."

Yeah? Smoke it.

[–] Daxtron2@startrek.website 2 points 5 months ago (1 children)

Lemmy.world and bandwagoning on a sensitive topic that they know nothing about? Classic combo.

[–] mindbleach@sh.itjust.works 2 points 5 months ago (1 children)

You'd figure "CSAM" was clear enough. You'd really figure. But apparently we could specify "PECR" for "photographic evidence of child rape" and people would still insist "he drew PECR!" Nope. Can't. Try again.

load more comments (1 replies)
load more comments (36 replies)
[–] Obonga@feddit.de 25 points 5 months ago (4 children)

13.000 images are generated relatively fast. My PC needs like 5 seconds for a picture with SD(depending on settings of course). So not even a day.

Also, if pedos would only create their own shit to fap to i would consider this a win.

load more comments (4 replies)
[–] SuperSpruce@lemmy.zip 21 points 5 months ago (1 children)

70 years for... Generating AI CSAM? So that's apparently worse than actually raping multiple children?

[–] Onihikage@beehaw.org 5 points 5 months ago (2 children)

He did more than generate it, he also sent some of it to a minor on Instagram, probably intending to get some real CSAM, or worse. For that, spending the next 70 years away from both children and computers seems appropriate to me.

load more comments (2 replies)
[–] stevedidwhat_infosec@infosec.pub 21 points 5 months ago (1 children)

Sensitive topic - obviously.

However these guard rail laws, and “won’t someone think about the children” cases are a reeeeally easy way for the government to remove more power from the people.

However, I believe if handled correctly, banning this sort of thing is absolutely necessary to combat the mental illness that is pedophilia.

[–] laughterlaughter@lemmy.world 8 points 5 months ago (2 children)

I don't condone child sexual abuse, and I'm definitely not a pedo (gosh, I can't believe I have to state this.)

But how does banning AI generated material help combating a mental illness? The mental illness will still be there, with or without images...

[–] Leg@lemmy.world 4 points 5 months ago (1 children)

There's something to be said about making it as difficult as possible to enable the behavior. Though that does run the risk of a particularly frustrated individual doing something despicable to an actual child. I don't exactly have the data on how all this plays out, and frankly I don't want to be the one to look into it. Society isn't particularly equipped to handle an issue like this though, focusing on stigma alone to kinda try to shove it under the rug.

[–] laughterlaughter@lemmy.world 3 points 5 months ago

Your second sentence is exactly what I was thinking of. The big issue with pedophilia is the fact that kids can be easily manipulated (or forced!) to do heinous acts. Otherwise, what's the difference with regular porn and topics about prisoners, slavery, necrophilia, etc? Would we say that people who consume rape fantasy porn will go out and rape? If a dude who is sexually attracted to women is not raping women left and right every day all year round, you know, because he knows it's wrong, if we're not labeling every heterosexual male as creeps, then why would this be different with other kinds of attractions?

But anyway. I'm not saying anything that hasn't been discussed in the past (I'm sure.) I'm just glad I don't have that condition (or anything similar, like attracted to volcanoes), otherwise life would definitely suck.

[–] stevedidwhat_infosec@infosec.pub 2 points 5 months ago* (last edited 5 months ago) (7 children)

Mainly it’s a problem of enabling the problem as others have mentioned.

It’s not a solution, per se. It doesn’t solve something specifically- but it doesn’t have to be. It’s about making it less accessible, harsher consequences, and so on to put more pressure on not continuing to participate in the activity. Ultimately it boils down to mental health and trauma. Pedophilia is a paraphilic disorder at the end of the day.

load more comments (7 replies)
[–] Darkrai@kbin.social 10 points 5 months ago (5 children)

It seems weird that the AI companies aren't being held responsible too.

[–] bjorney@lemmy.ca 19 points 5 months ago

It's open source code that someone ran on their own computer, it's not like he used paid OpenAI credits to generate the image.

It also would set a bad precedent - it would be like charging Solomons & Fryhle because someone used their (absolutely ubiquitous) organic chemistry textbook to create methamphetamine

[–] Demigodrick@lemmy.zip 7 points 5 months ago (1 children)

Well the American way is not to hold the company accountable, I.e. school shootings, so yeah.

[–] possiblylinux127@lemmy.zip 3 points 5 months ago (1 children)

I'm pretty sure you can't hold a school liable for a school shooting

[–] stanleytweedle@lemmy.world 13 points 5 months ago* (last edited 5 months ago) (8 children)

I think they were referring to the firearm manufacturer and\or seller.

load more comments (8 replies)
[–] JokeDeity@lemm.ee 6 points 5 months ago

Just to be clear, you guys think that any company that produces anything that ends up used in a crime should have criminal charges for making the product? Yeah, makes about as much sense as anything these days.

[–] Reddfugee42@lemmy.world 5 points 5 months ago

Was Kodak ever held responsible for original CSAM?

[–] jonne@infosec.pub 4 points 5 months ago

I think stable diffusion is an open source AI you can run on your own computer, so I don't see how the developers should be held responsible for that.

[–] HelixDab2@lemm.ee 9 points 5 months ago (1 children)

The basis of making CSAM illegal was that minors are harmed during the production of the material. Prior to CG, the only way to produce pornographic images involving minors was to use real, flesh-and-blood minors. But if no minors are harmed to create CSAM, then what is the basis for making that CSAM illegal?

Think of it this way: if I make a pencil drawing of a minor being sexually abused, should that be treated as though it is a criminal act? What if it's just stick figures, and I've labeled one as being a minor, and the others as being adults? What if I produce real pornography using real adults, but used actors that appear to be underage, and I tell everyone that the actors were all underage so that people believe it's CSAM?

It seems to me that, rationally, things like this should only be illegal when real people are being harmed, and that when there is no harm, it should not be illegal. You can make an entirely reasonable argument that pornographic images created using a real person as the basis does cause harm to the person being so depicted. But if it's not any real person?

This seems like a very bad path to head down.

load more comments (1 replies)
load more comments
view more: next ›