What Is a Deepfake Video?

In this post, I define a deepfake video as any audiovisual content that has been manipulated to impersonate another person. This manipulation can occur in the visual domain, such as face swapping, reenactment, or model-based synthesis; or in the audio domain, via text-to-speech (TTS) voice cloning or voice conversion.

While deepfakes can be entertaining or artistic, the real concern arises when they’re used for impersonation, misinformation, or character assassination. This raises an urgent need for detection methods that are both effective and generalizable.


A Quick Guide: What You’ll Find in This Post

This post is both a literature review and a personal research narrative. If you’re new to deepfake detection or want to skip to what’s most relevant, here’s a quick roadmap:


Unimodal Deepfake Detection

Video-Based Detection

Early deepfake detectors emerged from digital forensics, targeting signals that generative models typically fail to capture—like physiological cues and geometric constraints.

Examples include:

As deepfakes got better, frame-level forensic detectors appeared, followed by models analyzing motion across frames, such as optical flow or temporal consistency.


Audio-Based Detection

Audio deepfakes introduce their own telltale traces, either in signal properties or in the way they’re generated.

Approaches include:


Temporal Modeling in Deepfake Detection

Another key axis for classification is temporal modeling. Even high-quality fakes often fail to maintain consistency across frames.

Temporal methods fall into two groups:

  • Artifact-focused: Detect generation noise or device-specific patterns.
  • Identity-focused: Track how personal characteristics evolve over time.


Artifact-Focused Temporal Modeling

These models detect inconsistencies from compression or camera-specific signals:

A standout method is TI²Net, which compares face-identity embeddings over time. Using a recurrent model and triplet loss, it learns to spot identity fluctuations between frames, something that shouldn’t happen in real footage.


Identity-Focused Temporal Modeling

These methods track stable identity traits and compare them across time or against a reference.

These models often work best with speaker-specific reference videos and build strong personalized detectors.


Revisiting Audio-Visual Deepfake Detection

Let’s now focus on the intersection of audio and video — the most promising and complex area in detection (in my opinion).

Audio-visual deepfake detectors leverage the synchrony between speech and visual appearance. Depending on how features are extracted and compared, they fall into three broad categories:


1. Synchronization-Based Methods (Low-Level)

These methods focus purely on timing: does the mouth move when the voice says something?

They’re fast but don’t model speaker traits or expression dynamics.


2. Fusion-Based Methods (Low to High-Level)

These models fuse audio and video features into a joint representation using neural networks, usually guided by contrastive loss.

Low-level fusion (timing + surface-level correlation):

Mid- to high-level fusion (emotion, identity, semantics):

Fusion models are powerful but less interpretable than synchronization models.


3. Cross-Modal Biometric Matching

Here, the goal is to determine: does this face belong to this voice?

These models work well in identity verification settings but may struggle with unknown speakers.


What Contrastive Learning Actually Learns

Contrastive loss is everywhere — but what does it actually align?

Depending on the dataset, architecture, and training signals, the model might be learning any of the following:

  • Emotional consistency (e.g., smiles with happy tone)
  • Identity-related traits (e.g., facial structure ↔ timbre)
  • Timing (e.g., lip closures and consonants)
  • Vocal tract constraints (e.g., shape → formants)

Most fusion-based approaches mentioned earlier implicitly assume at least one of these correlations, or try to capture them in a data-driven way. Some methods explicitly focus on the first three (emotion, identity, timing), while others attempt to unpack the final one: anatomical alignment.

For example:

These ideas were a major inspiration for my own experiments. Even though my approach didn’t work out as hoped, it shaped how I think about what contrastive learning might actually be encoding, and what makes some of these signals easier to fake than others.


Reflections on My Own Experiments


My Hypothesis

Inspired by the articulatory literature, I tried to learn a joint representation from audio and visual motion features: deltas of audio features extracted with OpenSmile and optical flow.

I thought: instead of high-level semantics or identity, maybe we can just track movement synchrony — speech-induced motion — between lips and voice.

What Went Wrong

The model didn’t learn a joint space. Each modality formed its own subspace, and synchronization cues weren’t shared. The representation was not truly joint, likely because the design lacked an objective to encourage mutual dependence or cross-modal supervision. Therefore, there was no incentive to use one modality to help interpret the other.

Further down the path, I also experimented with two additional tasks:

  1. Cross-modal reconstruction: using audio to reconstruct video features, or vice versa
  2. Motion-based segment matching: determining whether a given audio snippet corresponds to a visual motion segment

Both attempts failed, likely due to the limitations of using a small, general-purpose audio-visual dataset. For example, the same event label (like “bell ring”) can appear with vastly different sounds and visual contexts, making it hard for the model to learn consistent cross-modal correspondences.

In hindsight, this line of work is more aligned with audio-visual event localization, where aligning semantics across modalities is already known to be challenging.

Moreover, cross-modal reconstruction may simply be too difficult on general datasets. While audio and visual motion in speech are tightly correlated due to shared underlying articulatory dynamics, general events lack that structure. Learning robust semantic representations in this setting would likely require much larger datasets with many instances per class — otherwise, the model risks overfitting or collapsing into shortcut features.


Ideas I Didn’t Pursue (but Still Like)

Either of them requires pre-training on a large dataset.

1. Speaking Style Modeling

Instead of phoneme-level synchrony, model speaking style over longer clips (5 to 10 seconds).

  • Temporal rhythm, prosody, gesture intensity
  • Contrastive learning could group similar styles together
  • May reveal differences that deepfakes can’t mimic

2. Vocal Tract Biometrics

Inspired by SAFARI, I imagined disentangling speaker physiology from speech content:

  • Stable: vocal tract shape, articulation range
  • Variable: phoneme identity, intonation

If we can extract a constant vocal tract representation, we could match that across time, even as phonemes change. This would be extremely hard for generative models to fake convincingly.


Final Thoughts

The central challenge in this arms race is identifying an asymmetry that gives defenders an advantage. If both the attacker and defender are training models, what gives the defender more leverage?

One promising answer lies in identity-based detection, where defenders can provide authentic reference videos at test time. In this setting, the defender may possess more data about the true subject than the attacker, enabling stronger verification.

Another, perhaps more fundamental, advantage lies in physiology and physical constraints. The human vocal tract, facial musculature, and speaking behavior are highly complex and deeply individualized — too intricate to be fully modeled by current generative systems. Yet these physiological signatures may be simpler to verify than to convincingly synthesize.

Until policy, law, and society catch up, technical defenses are our first line. Developing ensemble systems, some modeling artifacts, others modeling synchrony or biometrics, offer layered protection.

And they serve another role: helping people prove that a real video wasn’t fake, or vice versa. In a world where doubt can be weaponized, robust detection isn’t just technical infrastructure — it’s social infrastructure.