If AI and deep fakes can listen to a video or audio of a person and then are able to successfully reproduce such person, what does this entail for trials?

It used to be that recording audio or video would give strong information which often would weigh more than witnesses, but soon enough perfect forgery could enter the courtroom just as it’s doing in social media (where you’re not sworn to tell the truth, though the consequences are real)

I know fake information is a problem everywhere, but I started wondering what will happen when it creeps in testimonies.

How will we defend ourselves, while still using real videos or audios as proof? Or are we just doomed?

  • @[email protected]
    link
    fedilink
    English
    66 months ago

    Maybe each camera could have a unique private key that it could use to watermark keyframes with a hash of the frames themselves.

    • @[email protected]
      link
      fedilink
      56 months ago

      Usually I see non-technical people throw ideas like this and they’re stupid, but I’ve been thinking about this for a few minutes and it’s actually kinda smart

    • @[email protected]
      link
      fedilink
      English
      46 months ago

      I think that’s exactly how it’s going to work - you can’t force all ‘fake’ sources to have signatures- it’s too easy to make one without one for malicious reasons. Instead you have to create trusted sources of real images. Much easier and more secure

    • SavvyWolf
      link
      fedilink
      English
      56 months ago

      How would you prove that the camera itself is real, is the only device with access to the private key and isn’t falsifying it’s video feed?

      • @[email protected]
        link
        fedilink
        English
        16 months ago

        The sort of case I was thinking of is if different parties present different versions of an image or video and you want to establish which version is altered and which is original.