If AI and deep fakes can listen to a video or audio of a person and then are able to successfully reproduce such person, what does this entail for trials?

It used to be that recording audio or video would give strong information which often would weigh more than witnesses, but soon enough perfect forgery could enter the courtroom just as it’s doing in social media (where you’re not sworn to tell the truth, though the consequences are real)

I know fake information is a problem everywhere, but I started wondering what will happen when it creeps in testimonies.

How will we defend ourselves, while still using real videos or audios as proof? Or are we just doomed?

  • SavvyWolf
    link
    fedilink
    English
    59 hours ago

    How would you prove that the camera itself is real, is the only device with access to the private key and isn’t falsifying it’s video feed?

    • @AbouBenAdhem
      link
      English
      18 hours ago

      The sort of case I was thinking of is if different parties present different versions of an image or video and you want to establish which version is altered and which is original.

      • SavvyWolf
        link
        fedilink
        English
        16 hours ago

        You still have the same problem though. You can produce a camera in court and reject one of the images, but you still need to prove that the camera wasn’t tampered with and it was the one at the scene of the crime.

        • @LesserAbe
          link
          25 hours ago

          Leica has one camera that does this, and others are working on them. Just posted this link in another comment

          • SavvyWolf
            link
            fedilink
            English
            14 hours ago

            The camera can sign things however it wishes, but that doesn’t automatically make the camera trustworthy.

            In the same sense, I can sign any number of documents claiming to have seen a crime take place but that doesn’t make it sufficient evidence.