video-edit-mobileA picture can be worth a thousand words, or nothing at all. Deepfakes—highly realistic and personalized, but totally falsified, video or audio tapes—are changing the value and reliability of video and audio evidence.

"Deepfakes" is the name for highly realistic, falsified imagery and sound recordings; they are digitized and personalized impersonations. Deepfakes are made by using AI-based facial and audio recognition and reconstruction technology; AI algorithms are used to predict facial (mouth, jaw, eye, etc.) movements as well as vocal sounds.

Deepfake videos are created by taking an image or video clip of an individual—doing or saying anything at all. Through careful dissection and disassembly of each pixel of an image, or the dissection of a short audio clip of a voice, and using AI predictive technology, the software can create a doppelgänger doing something entirely different than the original image or sound clip. Indeed, what the doppelgänger may be doing on the fake tape is limited only by the imagination of the creator.

You don't have to be a technologist to access off-the-shelf software to create deepfakes, more and more of it is available free online. AI machine learning enables their creation. Just as the brain learns to recognize and identify a person by repeated exposure, the creator of a deepfake simply gathers a data set consisting of a number of images (still photographs or videoclips) of a target person (say a well-known political figure), and feeds it into an AI enabled program. That program compares the target image to a base image (say, an actor pretending to do and say something the politician would never say or do; or a video from a security camera showing the face of an actual perpetrator). The software uses machine learning to iteratively compare the images and effectively create a substitute of the target image onto the base image. The software learns from those images how the face appears in a number of different positions and is able to mimic those positions on the base image. One of the earliest uses of deepfakes was altering pornographic images with the faces of well known actresses.

Technology and techniques are being developed to detect deepfakes—either by recognizing a voice or image as falsified. Some of these efforts seem quite basic and manual: having humans do careful clip by clip analysis of suspect images or sound recordings and flagging potential errors that reveal them as fake. But there are also computerized pixel comparisons that can be done: taking a base image and analyzing it for missing, condensed or defective bits.

Technology is also being developed to tag new video and audio clips with indicia of authenticity. This new technology will be chasing improvements in the deepfake technology itself and provides only a partial solution. In addition, any certification of authenticity is primarily useful for new audio and video—while immediate harm may come from allegedly "discovered" and "old" audio and video clips created long before deepfakes came onto the scene.

How many times have we all heard "let's go to the tape," "seeing is believing," a "picture is worth a thousand words"? We are used to accepting what we see with our eyes and hear with our ears as "truth," and for that reason, in legal settings, audio and videotapes can be powerful evidence. A good quality audio or videotape can show who said what to whom, who was present in a specific location and who took a particular action. Audio and videotapes are used by the parties in myriad legal proceedings as well as prosecutors, judges and juries.

In a criminal case, a videotape of events can be the smoking gun that makes the difference between an Indictment or terminated investigation, a plea or going to trial, a conviction or acquittal. Imagine a videotape that shows the defendant holding a gun in a vestibule at the time that a murder occurred in that location; or buying a quantity of drugs in a location where a known drug sale took place. In a civil setting, imagine a videotape that shows that a person used a racial epithet or engaged in an inappropriate act at the office. If these visual images of people are fakes, they can result in enormous harm, including the conviction and incarceration of the wrong person.

Of course deepfakes outside of legal proceedings can also do great harm: fake tapes of political candidates saying or doing things they never said or did in order to influence the outcome of an election is an evil we should all be concerned about. But there are also national security issues: deepfakes being used to show governmental officials apparently taking military actions (for instance, authorizing a military strike) that they never took.

Deepfakes are changing our ability to rely on audio and video evidence: What we assume is truth might in fact be fiction.

What does the emergence of deepfakes mean for lawyers and judges? In 2019, it means there needs to be an awareness that they exist and to anticipate challenges (both well founded and opportunistic) to audio and videotape authenticity. We can expect the development of a cottage industry of people who purport to be able to tell a fake from an authentic tape, and to have a new series of experts in the courtroom. We may need to develop protocols for clear chain of custody histories for audio and videotapes; these histories will record the metadata demonstrating that a tape came off a particular server and was transmitted in a controlled manner.

In criminal cases, it means that defendants may argue that even if they appear on tape, "it wasn't me"; in cases in which the public is paying for the defense and associated costs, judges will be asked to approve costs of testing authenticity. This could become expensive and until technology has advanced, and not be guaranteed to result in a clear answer. Judges will then be faced with questions as to whether to allow a defendant to argue that a deepfake is the reason why the jury sees his face in the tape, in an effort to raise a reasonable doubt.

In civil cases, there will undoubtedly be Daubert and Frye challenges to the expertise of those individuals proffered as experts in differentiating real from fake; and the financial burden of undertaking that exercise may end up becoming a cost certain parties are unable to bear. An inability to effectively challenge a suspected or actual fake tape could make the difference between winning and losing a case.

We are entering a new era in which what we see cannot be equated with what we should believe. Lawyers and judges will need to keep abreast of developments in this area in order to confront challenges.

Katherine B. Forrest is a partner in Cravath, Swaine & Moore's litigation department. She most recently served as a U.S. District Judge for the Southern District of New York and was the former Deputy Assistant Attorney General in the Antitrust Division of the U.S. Department of Justice.