Video: WIRED Staff; Getty Images

The Thorny Art of Deepfake Labeling

Warning labels on AI-generated media give viewers little context. Artists and human rights advocates have forged a more effective—and creative—path.

Last week, the Republican National Committee put out a video advertisement against Biden, which featured a small disclaimer in the top left of the frame: “Built entirely with AI imagery.” Critics questioned the diminished size of the disclaimer and suggested its limited value, particularly because the ad marks the first substantive use of AI in political attack advertising. As AI-generated media become more mainstream, many have argued that text-based labels, captions, and watermarks are crucial for transparency.

But do these labels actually work? Maybe not.

For a label to work, it needs to be legible. Is the text big enough to read? Are the words accessible? It should also provide audiences with meaningful context on how the media has been created and used. And in the best cases, it also discloses intent: Why has this piece of media been put into the world?

Journalism, documentary media, industry, and scientific publications have long relied on disclosures to provide audiences and users with the necessary context. Journalistic and documentary films generally use overlay text to cite sources. Warning labels and tags are ubiquitous on manufactured goods, foods, and drugs. In scientific reporting, it’s essential to disclose how data and analysis were captured. But labeling synthetic media, AI-generated content, and deepfakes is often seen as an unwelcome burden, especially on social media platforms. It’s a slapped-on afterthought. A boring compliance in an age of mis/disinformation.

As such, many existing AI media disclosure practices, like watermarks and labels, can be easily removed. Even when they’re there, audience members’ eyes—now trained on rapid-fire visual input—seem to unsee watermarks and disclosures. For example, in September 2019, the well-known Italian satirical TV show Striscia la Notizia posted a low-fidelity face-swap video of former prime minister Matteo Renzi sitting at a desk insulting his then coalition partner Matteo Salvini with exaggerated hand gestures on social media. Despite a Striscia watermark and a clear text-based disclaimer, according to deepfakes researcher Henry Adjer, some viewers believed the video was genuine.

This is called context shift: Once any piece of media, even labeled and watermarked, is distributed across politicized and closed social media groups, its creators lose control of how it is framed, interpreted, and shared. As we found in a joint research study between Witness and MIT, when satire mixes with deepfakes it often creates confusion, as in the case of this Striscia video. These sorts of simple text-based labels can create the additional misconception that anything that doesn’t have a label is not manipulated, when in reality, that may not be true.

Technologists are working on ways to quickly and accurately trace the origins of synthetic media, like cryptographic provenance and detailed file metadata. When it comes to alternative labeling methods, artists and human rights activists are offering promising new ways to better identify this kind of content by reframing labeling as a creative act rather than an add-on.

When a disclosure is baked into the media itself, it can’t be removed, and it can actually be used as a tool to push audiences to understand how a piece of media was created and why. For example, in David France’s documentary Welcome to Chechnya, vulnerable interviewees were digitally disguised with the help of inventive synthetic media tools like those used to create deepfakes. In addition, subtle halos appeared around their faces, a clue for viewers that the images they were watching had been manipulated, and that these subjects were taking an immense risk in sharing their stories. And in Kendrick Lamar’s 2022 music video, “The Heart Part 5,” the directors used deepfake technology to transform Lamar’s face into both deceased and living celebrities such as Will Smith, O. J. Simpson, and Kobe Bryant. This use of technology is written directly into the lyrics of the song and choreography, like when Lamar uses his hand to swipe over his face, clearly indicating a deepfake edit. The resulting video is a meta-commentary on deepfakes themselves.

Activists’ and artists’ new takes on disclosure, like these, introduce new ways of seeing. Viewers of France’s documentary and Lamar’s music video are respectfully offered the visual language, vocabulary, and context to connect with and understand what they’re consuming. Both videos transcend easy answers and give audiences space to interpret for themselves. By contrast, despite their text-based labels, the Biden ad and the Italian satirical video fail to bring audiences into “the know” and leave them wondering, “Is this real or fake?”

As creators work to develop more detailed frameworks for deepfake and AI disclosure, disciplines and modes like accessibility theory, interactive storytelling, TikTok, footnoting practices, and museum image description guidelines all have useful tools to offer. In the art project Alt-Text as Poetry, audiences are encouraged to draft alt-text descriptions of images for visually impaired audiences that are poetic rather than perfunctory. Just like artistic disclosures, alt-text helps explain—or disclose—contextual information, ideally in a creative way. The artists explain that they approach access “generously, centering disability culture, rather than focusing on compliance.” On TikTok, tags on videos and hashtags in captions provide insights into how users create videos and interact with each other through remixes, duets, snappy editing, AI effects, and filters. As a result, the app’s labeling system becomes an integral and fun part of the platform's engagement mechanism, showcasing the creative potential and social benefits of revealing the production process.

These context-driven labeling models engage users while making clear how these images have been created and manipulated. When creators go beyond the bare minimum of compliance, they can produce work that is more innovative and more principled. Art can illuminate.


WIRED Opinion publishes articles by outside contributors representing a wide range of viewpoints. Read more opinions here, and see our submission guidelines here. Submit an op-ed at ideas@wired.com.