THE SENATE UNANIMOUSLY passed a bipartisan bill to provide recourse to victims of porn deepfakes — or sexually-explicit, non-consensual images created with artificial intelligence.

The legislation, called the Disrupt Explicit Forged Images and Non-Consensual Edits (DEFIANCE) Act — passed in Congress’ upper chamber on Tuesday.  The legislation has been led by Sens. Dick Durbin (D-Ill.) and Lindsey Graham (R-S.C.), as well as Rep. Alexandria Ocasio-Cortez (D-N.Y.) in the House.

The legislation would amend the Violence Against Women Act (VAWA) to allow people to sue those who produce, distribute, or receive the deepfake pornography, if they “knew or recklessly disregarded” the fact that the victim did not consent to those images.

  • @Contravariant
    link
    101 month ago

    The worrying aspect of these laws are always that they focus too much on the method. This law claims to be about preventing a particular new technology, but then goes on to apply to all software.

    And frankly if you need a clause about how someone is making fake pornography of someone then something is off. Something shouldn’t be illegal simply because it is easy.

    Deepfakes shouldn’t be any more or less illegal than photos made of a doppelgänger or an extremely photorealistic painting (and does photorealism even matter? To the victims, I mean.). A good law should explain why those actions are illegal and when and not just restrict itself to applying solely to ‘technology’ and say oh if it only restricts technology then we should be all right.

    • @postmateDumbass
      link
      121 month ago

      I am not ok with making art a crime.

      Do not give away basic human rights because of emotional appeals.

      If sexual blackmail is the problem, prosecute sexual blackmail.

      • Todd Bonzalez
        link
        fedilink
        21 month ago

        Where do you draw the line at “art”. If nonconsensual pornography is acceptable to you, how about child porn? Where does art end and crime begin?

        • @postmateDumbass
          link
          21 month ago

          Use existing rules and laws for that.

          CSAM is illegal across all mediums.

          • Todd Bonzalez
            link
            fedilink
            6
            edit-2
            1 month ago

            Okay, and we already have laws against nonconsensual pornography. But here’s the thing: those laws could easily be argued to only count for actual photographs or videos, in that many of these laws were written to tackle “revenge porn” and not “fake porn”, so someone making nonconsensual porn with AI could get charges dropped on the grounds that the porn wasn’t legitimate, or didn’t meet the definition of pornography starring the victim as defined by existing laws (because those laws were written before AI made this a mainstream problem).

            Making laws that explicitly list things like photoshop and AI faceswap or “roop” tools that can create artificial nonconsensual pornography lays the groundwork for consequences for people who sexually abuse others in this way.

            And let’s be realistic, if you use AI to make nonconsensual porn for personal purposes, you may be a creep, but you also aren’t going to get caught if you keep it to yourself. You also aren’t going to harm the person you’re making porn of if you’re not distributing the content you’re creating. You still shouldn’t do it, but it’s practically legal if you simply keep your weird shit to yourself and don’t publicly victimize someone with it. The people who will get caught are going to be in the latter category, and I have no problem seeing the book thrown at them.

            • @postmateDumbass
              link
              01 month ago

              Just as you point out revenge porn laws have a problem because they miss the mark covering fake porn, outlawing the methods amd mediums of art ued today is also going to miss the mark. Aside from the oppression issues, new methods will come along and circumvemt the targeted laws.

              Just need to write laws that target the crime, here it would seem to be a combination of defamation, privacy, and assault in play for using sexuality or nudity to harm an individual and/or their reputation. Without harm, what crime is there?

    • @[email protected]
      link
      fedilink
      31 month ago

      You can tell when a painting is a painting. Even it’s photorealistic. You know a person created it and that it’s fiction. Often these are hung in galleries where people expect to see art.

      Deep fakes exist to fool people into thinking someone did something(like pornography) when they didn’t…usually with the intention of causing harm to their reputation. That’s already illegal due to defamation laws, so really it’s just an extention of those combined with revenge porn laws.

      The reason they have to include the type of tech in the law is because that tech made it possible for unskilled bad actors to get on it…therefore there’ll be more people committing these types of crimes against others. It’s a good thing they’re addressing this issue.

      • @Contravariant
        link
        2
        edit-2
        1 month ago

        The reason they have to include the type of tech in the law is because that tech made it possible for unskilled bad actors to get on it

        Yeah, and that’s the part I don’t like. If you can’t define why it’s bad without taking into account the skill level of the criminal then I’m not convinced it’s bad.

        As you point out defamation is already illegal and deliberately spreading false information about someone with the intent to harm their reputation is obviously wrong and way easier to define.

        And is that not why you consider a painting less ‘bad’? Because it couldn’t be misconstrued as evidence? Note that the act explicitly says a digital forgery should be considered a forgery even when it’s made abundantly clear that it’s not authentic.

        • @[email protected]
          link
          fedilink
          51 month ago

          Look, I’m a professional artist. The general rule is you have to change something 15% to 30%(depending on location) for it to not come into violation of copyright laws. That’s why you see satirical depictions of brands in cartoons and such.

          This new law has to take into consideration art laws, defamation laws, revenge porn laws, slander laws, and the right for a person to own their likeness.

          It is absolutely necessary to reign this in before serious harm is done to someone. The point of writing a law to address this specific issue is because for the law actually be effective, it must be written to address the specific problem this technology presents. I listed the other laws to show its consistent with ones we already have. There’s nothing wrong with adding in another to protect people.

          As for the unskilled part, the point of that is a skilled person creating deepfake porn by hand, frame by frame should get in as much trouble as an unskilled person using ai. The AI is just going to make it so more unethical people are making this crap…so more if it will exist. That’s a problem that needs addressing.

          You have a nice day now.

    • @ByteJunk
      link
      21 month ago

      There are stark differences between the scenarios you’re presenting, but going to the core of your point, is it even legal to paint a photorealistic nude?

      I don’t know of any court cases about this specific subject, but I remember when Rush painted Tiger Woods (“The masters at Augusta”), he was sued.

      He got away with not having to pay money to Tiger Woods, but partly because it’s a stylized painting and it pushed towards first amendment rights. This wouldn’t work in a photorealistic depiction, so it seems highly unlikely that such a painting would be OK…

      • @Contravariant
        link
        -21 month ago

        Fair, but then this law serves no purpose. The thing it was designed to prevent was already illegal.