Our Resources
/
Research and Publications

Building a Glossary for Synthetic Media Transparency Methods

Part 1: Indirect Disclosure

$hero_image['alt']

According to an Axios poll, a third of Americans say AI will reduce their trust in the results of the 2024 U.S. election. For Partnership on AI (PAI) and our community of organizations building, creating, and distributing synthetic media, this raises important questions: How can everyday citizens know if the media they see and hear has been created or modified by AI and, more importantly, is misleading? And what can organizations do to improve transparency?

There is no perfect answer. However, those building, creating, and distributing synthetic media can act to support better explanations of how content has been created and edited. They have a responsibility, when appropriate, to make clear to audiences that content is AI-generated or AI-modified (and any uncertainty about such judgments).

While many disparate efforts have emerged to help audiences navigate an increasingly synthetic information environment, largely by providing context about content, the community has not aligned on which combination of tactics to implement, when to share insights with audiences, and how they can evaluate their efficacy in supporting trustworthy content.

The need for alignment is urgent. Policymakers—including the White House in its recent Executive Order—have begun to explore different techniques for technology, government, civil society, and media organizations, to provide audiences with greater insight into what they see and hear.

To improve alignment and understanding of synthetic media transparency, PAI has developed a series of resources. Part 1 of the series, focused on indirect disclosure, includes:

  1. A community-driven glossary defining the common technical methods that can provide insight into whether media is synthetic or not—what we describe as indirect disclosure methods;
  2. Proposed questions for evaluating indirect disclosure methods; and
  3. Initial guidance on how policymakers and those building, creating, and distributing synthetic media should implement these techniques.

The glossary below builds off of a seven-part workshop series with 18 organizations that are collaborators on PAI’s Responsible Practices for Synthetic Media. While the information in this document is provided by PAI and is not intended to reflect the view of any other organization, much of it has been informed through collaboration.

Different organizations use different terms to refer to the class of methods described below. However, we encourage the use of the term indirect disclosure to center these methods in how they contribute to the goal of broader AI transparency. While the disclosure taking place will not always translate to transparency for audiences, it does reveal details about content’s AI-generation or AI-modification to entities involved in content development, creation, and distribution. Such information can be used to support and underscore broader societal transparency goals. We will explore direct disclosure in future blog posts.

Graphic Notes: Synthetic Media Detection is set apart in the graphic since it is slightly different from other technical methods for synthetic media transparency; such detection derives signals from media and does not require any action taken to implant a signal in content, like indirect disclosure methods do. See below for more detail on the proactive and derived dynamics of these methods.

Glossary

We include methods in the glossary based on a combination of factors, including whether or not they are in use by those developing, creating, and distributing synthetic media, as well as input from our 18 Framework collaborators. Inclusion in the glossary does not equate to recommended use—rather, the Initial Recommendation section below offers initial guidance on the optimal combination of indirect disclosure methods, which we will further develop in future blogs. We also include key terms in the glossary to help support understanding of the methods in the graphic above.

Synthetic Media (or “Generative Media”)

Visual, auditory, or multimodal content that has been generated or modified (commonly via artificial intelligence). Such outputs are often highly realistic, would not be identifiable as synthetic to the average person, and may simulate artifacts, persons, or events.

Examples
Synthetic Media Transparency Methods

The umbrella term used to describe signals for conveying whether a piece of media is AI-generated or AI-modified. Such signals can either be indirect (not user facing) or direct (user facing). Indirect disclosure signals can support understanding of whether content has been AI-generated or AI-modified and, when appropriate, guide development of direct disclosures to audiences and end-users.

Examples

Indirect disclosures, direct disclosures, synthetic media detection

Indirect Disclosure

A signal for conveying whether a piece of media is AI-generated or AI-modified, based on information about a piece of content’s origin and/or its evolution; is not user facing. The “disclosure” that takes place is typically to entities involved in content development, creation, and distribution—but it can be used to inform direct, or audience/user-facing, disclosure.

Examples

Watermarks, fingerprints, metadata

Proactive Methods

  • At generation
  • Post-generation

Synthetic media transparency methods that rely upon an actor purposefully applying a signal that can then be identified by a third-party who is able to detect or interpret the signal. Notably, in an adversarial setting, bad actors will not leverage such techniques and may attempt to alter existing signals.

Can be classified further into two categories:

At generation

Signal is applied automatically by a media generation model at the moment of creation.

Post generation

Signal is applied after creation.

Examples

Watermarks, fingerprints, metadata

Derived Methods

Synthetic media transparency methods that determine the origin or evolution of media based on signals that do not rely upon a disclosure signal being applied by an actor.

Examples

Synthetic media detection

Direct Disclosure

A signal for conveying to users whether a piece of media is AI-generated or AI-modified; often informed by indirect disclosures.

Examples

Labels, content overlays

Watermarking 

  • Invisible
  • Visible

The proactive (at, or post, generation) insertion of modifications into a piece of content that can help support interpretations of how the content was generated and/or edited. Can come in two forms:

Invisible

Modifications made to a piece of content that are imperceptible to the human eye or ear. Can only be identified by a watermark detector (distinct from Synthetic Media Detection).

Visible

Modifications made to a piece of synthetic content that are detectable to the human eye or ear and do not require the use of a detector to interpret them.

Examples

Google’s SynthID (invisible), Meta AI’s Imagine (visible)

Watermark Key/Detector
A digital tool, similar to a password, that is required for detecting an invisible/ hidden watermark embedded in a piece of content. Can be shared broadly (open) or be restricted to select players (closed), or in between. This choice can affect the technical robustness of a watermark and its societal impact.

Fingerprinting 

  • Cryptographic hashing
  • Perceptual hashing

The proactive (at, or post, generation) process by which a hash is generated for a piece of content for the purpose of identifying that content at a later date. Such hashes must be stored in a database in order to verify future content against the original. Unlike watermarking, this hash is not embedded in the content file itself. Also known as “hashing and matching” or “hashing and logging.” Can come in two forms:

Cryptographic hashing

An exact-match form of hashing where the hash for a piece of synthetic content will not match if the content has been modified in any way.

Perceptual hashing

A probabilistic-match form of hashing where the hash for a piece of synthetic content is resilient to minor perturbations (i.e., will still match with minor changes).

Example

YouTube’s Content ID

Metadata 

  • Signed
  • Unsigned

Information about the origin, structure, and/or editing history of a piece of content that is proactively attached to the content itself.

Signed Metadata

Information that is proactively attached to the content itself and stored using secure encryption; a trusted/validated signer certificate is added post generation. State of the art methods leverage cryptographic signatures.

Unsigned Metadata

Information that is proactively attached to the content itself at generation but is not stored with secure encryption or validated with a trusted signer certificate, and potentially can be changed imperceptibly (weakening robustness).

Note

In policy and public discourse, metadata has sometimes been described as “media provenance” despite the fact that provenance can be a broader term describing all the methods under the umbrella of indirect disclosure. For example, several PAI Partners refer to indirect disclosure methods as provenance methods. It is our hope that the terminology expressed in this glossary helps align the field on nomenclature.

Examples

C2PA Standard (signed), IPTC Standard (unsigned)

Synthetic Media Detection

Methods that rely on detecting unintentionally added patterns/forensic cues differentiating synthetic media from non-synthetic media to determine the likelihood that a piece of content was AI-generated or AI-modified; such methods do not rely on theproactive addition of artifacts such as watermarks in content. Synthetic media detection is a derived transparency method.

Examples

Intel’s FakeCatcher, Google Jigsaw’s Assembler (no longer active)

Detector Access

  • Open
  • Closed

Whether for watermark detection or synthetic media detection, systems identifying synthetic media can either be shared broadly or restricted in their access, or somewhere in between.

Closed Detectors

Only available to a select number of organizations, minimizing the risk of adversarial exploitation at the expense of accessibility.

Open Detectors

Widely available, maximizing accessibility at the expense of increased risk of adversarial exploitation.

Note

Detector access does not necessarily have to be a binary choice between open and closed. PAI has conducted much work on the tradeoffs between open and closed access; see here, here, and here, all incorporating recommendations for “goldilocks” exposure between open and closed.