Watermark

As AI becomes more prolific, it's increasingly important for people and for platforms to be able to differentiate between generated content and human-created content.

  • For consumers, this helps you sort through the bombardment of online content with more confidence and agency.
  • For creators, this protects the rights of your content from reproduction and counterfeits.
  • For researchers and platform owners, this protects the quality of the inputs into your work.

Watermarks help us by adding or identifying tracers into text, images, sounds, and video when it's created digitally by a model.

[Note: For the sake of simplicity in terms, this will combine the two approaches of digital watermarking and content provenance into one term]

Watermark types

  • Overlay watermarks are visual symbols or text added superficially to the content as a post-processing technique. They can be added automatically by the generative engine, or by the platforms where the content is uploaded. Overlay watermarks are not integrated into the structure of the content and can therefore be easily removed.
  • Steganographic watermarks embed patterns into the content that are imperceptible to humans. These might include subtle changes to pixels in video and images, additional spaces in text, or slight note changes in audio files. While slightly better protected than overlay watermarks, these can be masked intentionally or by bad actors through small changes to the file such as adding a Gaussian blur.
  • Machine learning watermarks use a new approach that relies on machine learning models to add a distinct key to the content that can only be read by other models. This is the process Google is using in their SynthID approach. They are stronger than the first two types, but can be degraded as the image is modified.
  • Statistical watermarks are random patterns injected into the image, audio, or text by the generator itself. While still impacting the content at the surface level (and not the foundational structure of the content), the randomness of this approach makes it far more difficult to crack or mask, especially for the average use.

Compared to content provenance

This alternative approach embeds data into the origin (or provenance) of the content itself, creating a digital fingerprint on the content's metadata. Content provenance requires the cooperation of content generators and digital platforms to create and enforce these standards.  

Many large tech platforms have signed onto a pledge to incorporate standards developed by the Coalition for Content Provenance and Authenticity (C2PA), which defines a common format that cannot be tampered with without leaving traces of manipulation. This data could then be read by any platform that adopts the protocol, including the full history of modifications.

The Content Credentials Verifier allows you to upload content and see what watermarks are embedded in its metadata

Regulatory landscape

Early steps are being taken by many governments to enforce watermarks in both scenarios.

  • China requires source-generated watermarks, as well as the footprint pattern of including metadata in the filename when downloaded.
  • The European Union's AI Act imposes similar standards regarding labeling generated content, and also imposes requirements on foundation models to ensure transparency and compliance with copyright law.
  • In the United States, a 2023 Executive Order established standards for watermarking and regulation, but requires action by Congress to codify these standards for enforcement.

The implications of this regulatory tapestry remain to be seen. If you are working on a foundational model or generative tool, you can prepare by developing principles and standards to evaluate the product experience. If you are working on a platform, think through how to label this content in different scenarios.

There is not a conventional approach established that mitigates the conflicting risk to this issue, and most likely it will need to be solved through a combination of regulation and source code. Additional innovation is being pursued through other means, like MIT's work on PhotoGuard, which essentially blocks generative tools from manipulating certain aspects of an image.

Designers can lean in to help establish human-centered principles and conventions, and carefully construct copy and caveats to help users be as informed as possible.

If you want to track the progress of this issue, here are some resources to get started:

Articles and Explainers
Policies, Protocols, and Tools

Design considerations

  • Make watermark visibility user-centered. Do not assume one level of visibility fits all cases. Consumers may need clear overlays or labels, while creators and researchers may prefer invisible or metadata-level tracers. Match visibility to the audience and context.
  • Support both imperceptible and visible watermarks. Visible labels deter casual misuse and clarify origin. Invisible tracers, like embedded hashes or model fingerprints, provide forensic accountability. Combining both gives broader protection across the content lifecycle.
  • Provide context when surfacing watermarks. Users need to know not just that provenance data exists but what it means. Include metadata such as source model, generation time, applied edits, and verified publisher when available.
  • Standardize placement across surfaces. Keep watermarks details in predictable locations, such as a “Content Info” panel, metadata view, or hover state. Consistency reinforces recognition and reduces fatigue.
  • Balance regulation and choice. Where watermarking or provenance disclosure is mandated, comply fully. Elsewhere, offer creators options for when and how attribution is applied, while preserving baseline protection for consumers.

Examples

Adobe's Content Credentials protocol has been adopted by TikTok and other platforms with a shared icon intended to be a universal indicator of synthetic content
TikTok became the first social media company to support the Content Credentials Metadata tags for all content created on their platform.
IA Writer offers an example of a tool proactively watermarking generated text, distinguishing it in grey from text actually authored by the user