If AI and deep fakes can listen to a video or audio of a person and then are able to successfully reproduce such person, what does this entail for trials?

It used to be that recording audio or video would give strong information which often would weigh more than witnesses, but soon enough perfect forgery could enter the courtroom just as it’s doing in social media (where you’re not sworn to tell the truth, though the consequences are real)

I know fake information is a problem everywhere, but I started wondering what will happen when it creeps in testimonies.

How will we defend ourselves, while still using real videos or audios as proof? Or are we just doomed?

  • Randomgal@lemmy.ca
    link
    fedilink
    arrow-up
    13
    ·
    27 days ago

    A bit dramatic imo. For most of legal history we didn’t actually have perfectly recorded video or audio, and while they are great tools at the present, they are still not the silver-bullet people would expect them to be at trial. (Think Trump and his cucks) Furthermore, most poor people try to avoid being recorded when doing crimes.

    It will probably mean that focus will shift to other kinds of evidence and evidence-gathering methods. But definitely not the end of law as we know it, far from it.

    • ConstipatedWatson@lemmy.worldOP
      link
      fedilink
      arrow-up
      2
      ·
      27 days ago

      Right, but anyone would like to not be in a video implying them in a crime, but I was wondering what would happen if fake videos of said person were to appear implicating a crime that actually did not take place