395
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
this post was submitted on 08 Dec 2023
395 points (93.2% liked)
Technology
60090 readers
2636 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 2 years ago
MODERATORS
Billy doesn’t need a nudify app to imagine Stacy naked. Not to mention, images of a naked 13 year old are illegal regardless.
Why are you pretending that "nudify apps" are produce ephemeral pictures equivalent to a mental image? They are most definitely not.
Underage teenagers already HAVE shared fake porn of their classmates. It being illegal doesn't stop them, and as fun as locking up a thirteen year old sounds (assuming they get caught, prosecuted, and convicted) that still leaves another kid traumatized.
So if illegality doesn’t stop things from happening… how exactly are you stopping these apps from being made?
Go after the people advertising those apps. Developers and advertisement agencies who say/intentionally imply "create naked pictures of people you know" should all be prosecuted.
Unlike photoshop or generic SD software, these apps have literally no legitimate reason to exist since the ONLY thing they facilitate is creating non-consensual pornography. Seems like something that would be very easy to criminalize.
So wait, we can’t criminalize the use, but if we criminalize the advertisement it fixes the situation?
You realize the exact same problem exists? There are plenty of tools with illegal uses, easily accessible online right now. Many on GitHub.
Just as the other people in this made up scenario don't need an app to imagine Scarlet Johansen naked. It's a null point.
I think most of this is irrelevant because the tool that is AI image generation is inherently hard to limit in this way and I think it will be so prevalent as to be hard to regulate. What I’m saying is: we should prepare for a future where fake nudes of literally anyone can be made easily and shared easily. It’s already too late. These tools, as was said earlier, already exist and are here. The only thing we can do is severely punish people who post the photos publicly. Sadly, we know how slow laws are to change. So in that light, we need to legislate based on long term impact instead of short term reactions.
And?… There’s a major difference between “a lookalike of a grown adult” and “ai generated child porn” as im sure you’re aware. At no point did anyone say child porn was going to be legal, until the person I was replying to brought it up as a strawman argument. ¯\_(ツ)_/¯