Rep. Joe Morelle, D.-N.Y., appeared with a New Jersey high school victim of nonconsensual sexually explicit deepfakes to discuss a bill stalled in the House.
Rep. Joe Morelle, D.-N.Y., appeared with a New Jersey high school victim of nonconsensual sexually explicit deepfakes to discuss a bill stalled in the House.
I really wonder whether this is the right move.
This girl, and many others, are victims and I don’t want to diminish that, but I for better or worse I just don’t see how legislation can resolve this.
Surely deepfakes will be just different enough to the subject to create reasonable doubt that it depicts the subject.
I wonder whether, as deep fakes become commonplace, people might be more willing to just ignore it like any other form of trolling.
I think you’re right if the goal is to stop them all together.
But what we can do is stop people from sending them around and saying that it’s true/actually the person.
Once they’ve turned it from a art project into a weapon, it should have similar consequences to “revenge porn.”
I would think this would be covered by libel, slander, defamation type laws. The crime is basically lying about a persons actions and character.
I think it doesn’t go far enough. Straight up, no one should be permitted to create or transmit the likeness of anyone [prior to, say, 20 years following their death] without their explicit, written permission. Make the fine $1,000,000 or 10% of the offender’s net worth, whichever is greater; same penalty and corporate revocation for any corporation involved. Everyone involved from the prompt writer to the work-for-hire people should be liable for the full penalty. I can’t think of a valid, non-entertainment (parody/humor), reason for non-consensual impersonation - and using it for humor or parody is a slippery slope to propaganda weaponization. There is no baby in this tub of bathwater.
Yeah, just like the FBI warnings on VHS tapes about massive fines and jail time stopped us from copying them in the 80s and 90s…
It’s not trolling it’s bullying. You need to think beyond this being about “porn”. This is a reputational attack that makes the victim more likely to be further victimised via date rape, stalking, murder. These things already happen based on rumours, deepfakes images/videos will only make it worse. The other problem is that it’s almost impossible to erase once it’s on the internet, so the victim will likely never be free of the trauma or danger as the images/videos resurface.
Trolling / bullying is just semantics, which I don’t think will help us very much.
I think the heightened risk of other crimes is… dubious. Is that conjecture?
Your position seems to be framed in the reality of several years ago, where if you saw a compromising video of someone it was likely real, while in 2024 the opposite is true.
Were headed towards a reality where someone can say “assistant, show me a deepfake of a fictitious person who looks a bit like that waitress at the Cafe getting double teamed by two black guys”. I don’t claim to know all the ethical considerations, but I do think that changing social norms are part of the picture.
I don’t have any authority to assert when anyone else should feel victimised. All I know is that in my own personal case, a few years ago I would’ve felt absolutely humiliated if someone saw a compromising video of me, but with the advent of deep fakes I just wouldn’t care very much. If someone claimed to have seen it I would ask them why they were watching it, and why in the world they would want to tell me about their proclivities.
My dude there are people out there thinking they’re in a relationship with Johnny fucking Depp because some Nigerian scammer sent them five badly photoshopped pictures. Step out of your bubble, maybe. This shit isn’t easy to spot for the vaaaaaast majority of people and why would this lie with the victim to sort of clear their name or hope that idiots realize it’s fake?
Especially with and around teenagers who can barely think further than their next meal?
Good lord.
WDYM “step out of your bubble”?
It’s not a question of being able to detect whether or not a video is fake. When deepfakes become so prevalent that everyone’s grandma understands that they’re prevalent, it won’t matter whether you can identify the video as fake.