- cross-posted to:
- [email protected]
- cross-posted to:
- [email protected]
A nightmare scenario previously only imagined by AI researchers, where AI image generators accidentally spit out non-consensual pornography of real people, is now reality.
A nightmare scenario previously only imagined by AI researchers, where AI image generators accidentally spit out non-consensual pornography of real people, is now reality.
But not over that line. If you draw lines and then these companies stay within them, what’s the problem?
People don’t have unique appearances.
Well, setting aside that there’s no law against being a total asshole, so like… We don’t have to make a bad behavior illegal in order to complain about it…
There’s the letter of the law, and there’s the intent. We start with a shared cultural attitude of how we should treat each other, and then we turn that into a quantifiable, objective rule that we can enforce through law.
We can try to make the law match our cultural attitudes as closely as possible, but there will always be gaps.
Now, I’ve got my own beef with how our IP and publicity laws work, and I’d like them to be more permissive in many ways. Much of IP law is exploitative, takes advantage of creators more than it protects them, and has lagged way behind where our social norms are these days.
But these ML companies aren’t interested in abiding by any social norms at all. Only paying lip service to current laws, which were written in a time before these “AI” services were even a possibility – skating by on technicalities, like a little brother poking the air 2 inches from your face and taunting “I’m not touching you! I’m not touching you!”