How to Detect Deepfakes: A Practical Guide to Spotting AI-Generated Media

In 2026, the barrier to creating convincing fake video, audio, and images has effectively collapsed. What once required a team of visual effects specialists and hundreds of thousands of dollars in computing resources can now be done by a single person with a consumer laptop and freely available software. The term "deepfake" -- a portmanteau of "deep learning" and "fake" -- has moved from a niche concern among AI researchers to a mainstream threat affecting elections, financial markets, personal reputations, and everyday trust in what we see and hear.

The consequences are not hypothetical. Deepfake audio has been used to impersonate CEOs and authorize fraudulent wire transfers. Fabricated video of political figures has circulated during election cycles to mislead voters. Synthetic pornography has been weaponized to harass and extort individuals. And as the technology improves, the window of time between a deepfake's release and its debunking continues to shrink in relevance -- the damage is done before the correction arrives.

This guide provides practical, actionable techniques for identifying deepfakes across video, audio, and images. It is written for journalists, researchers, educators, and anyone who wants to develop a sharper eye for synthetic media. No specialized technical background is required.

How Deepfakes Are Made: Understanding the Technology

Understanding how deepfakes are created helps you understand where their weaknesses lie. The two dominant approaches are generative adversarial networks (GANs) and diffusion models, and each leaves different artifacts.

GANs work by pitting two neural networks against each other. The generator creates synthetic content, and the discriminator evaluates whether that content looks real. Through thousands of iterations, the generator gets better at producing convincing fakes, and the discriminator gets better at spotting them. The output is a kind of evolutionary arms race within the model itself. GAN-generated faces often exhibit telltale symmetry issues, unusual skin textures, and background inconsistencies because the model optimizes for the face while treating the surrounding context as secondary.

Diffusion models, which power tools like Stable Diffusion and DALL-E, take a different approach. They start with pure noise and gradually refine it into a coherent image, guided by text prompts or reference images. These models tend to produce more natural-looking output than GANs but often struggle with fine structural details -- hands with the wrong number of fingers, text that does not quite resolve into real words, and objects with physically impossible geometry.

For video deepfakes, the most common technique is face-swapping, where a source face is mapped onto a target face in existing footage. The model must track facial movements, match lighting conditions frame by frame, and blend the edges of the swapped face into the original video. Each of these steps introduces potential artifacts that a careful observer can detect.

Voice cloning technology has also advanced dramatically. Modern text-to-speech systems can replicate a person's voice from as little as three seconds of sample audio. These systems capture pitch, cadence, and tonal qualities, but they often miss the subtle irregularities -- breathing patterns, vocal fry, micro-hesitations -- that characterize natural speech.

Visual Tells: What to Look for in Deepfake Video and Images

No single visual artifact is a guaranteed indicator of a deepfake, but several patterns appear frequently enough to serve as reliable signals.

Eyes and gaze. Deepfakes historically struggled with eyes. Look for unnatural reflections in the irises -- in authentic photos, the reflections in both eyes should be consistent because they are reflecting the same environment. Deepfakes sometimes produce different reflections in each eye or reflections that do not correspond to any plausible light source. Blinking patterns can also be irregular; early deepfakes rarely blinked at all, and while newer models have corrected this, the timing and naturalness of blinks can still appear off.

Teeth and mouth. The interior of the mouth is one of the hardest areas for deepfake models to render convincingly. Watch for teeth that appear blurred, fused together, or inconsistent between frames. The tongue may look flat or textureless. When the subject speaks, the synchronization between lip movements and the shape of the mouth for specific phonemes may not quite match, particularly for sounds like "M," "B," and "P" that require the lips to close fully.

Hair and edges. The boundary between a person's hair and the background is a persistent challenge for deepfake algorithms. Look for hair that appears unnaturally smooth, moves in a blocky fashion, or has a halo-like fringe where it meets the background. Individual strands are computationally expensive to render, so deepfake hair often looks slightly painted or lacks the fine detail visible in authentic high-resolution footage.

Skin texture and lighting. Deepfake skin can appear too smooth, almost waxy, particularly in areas like the forehead and cheeks. Pores, moles, and fine lines may be inconsistent or absent. Lighting is another common failure point -- the direction and quality of light on the face should be consistent with the lighting on the body and background. Look for shadows that fall in wrong directions or highlights that appear on the wrong side of the face.

Temporal inconsistencies. In video deepfakes, watch for flickering or shimmering around the edges of the face, particularly along the jawline and hairline. These artifacts are often visible when the subject turns their head quickly or when the lighting changes. Pausing the video and advancing frame by frame can reveal blending artifacts that are invisible at normal playback speed.

Background anomalies. Because deepfake models focus computational resources on the face, the background of the image or video may contain distortions. Look for warped lines, objects that appear to bend unnaturally near the subject's head, or background elements that shift inconsistently between frames.

Audio Deepfake Clues: Detecting Synthetic Speech

Audio deepfakes present a distinct set of detection challenges because there are fewer visual cues to rely on. However, synthetic speech has its own characteristic weaknesses.

Unnatural prosody. While voice cloning captures the basic tone and pitch of a voice, it often fails to reproduce the natural rhythm of speech. Listen for sentences that feel mechanically paced -- too even in tempo, with pauses that fall at algorithmically convenient points rather than where a human speaker would naturally breathe or hesitate.

Breathing and ambient sound. Real speech includes the sounds of breathing, lip smacking, and interaction with the acoustic environment. Deepfake audio often has an unnaturally clean quality, as if the speaker is in a perfect vacuum. Alternatively, the background noise may be inconsistent -- present in some segments and absent in others -- suggesting that different portions of audio were generated separately.

Emotional flatness. Cloned voices often struggle with emotional range. A speaker who should sound angry may sound merely louder. Someone delivering sad news may maintain an oddly neutral tone. These mismatches between content and delivery are worth noting.

Phoneme transitions. Pay attention to how the voice transitions between sounds. Natural speech involves complex articulatory movements that affect the character of adjacent sounds -- the "n" in "pan" sounds subtly different from the "n" in "ping" because of the surrounding vowels. Synthetic voices sometimes fail to reproduce these coarticulatory effects, resulting in speech that sounds subtly disconnected.

Spectral analysis. For those willing to use technical tools, viewing audio in a spectrogram can reveal anomalies invisible to the ear. Synthetic audio may show unnaturally smooth frequency distributions, abrupt discontinuities at edit points, or the absence of harmonic overtones that characterize natural voice production. Free tools like Audacity allow you to generate spectrograms from audio files with minimal effort.

Metadata Analysis: The Digital Paper Trail

Before analyzing the content of a suspected deepfake, examine its metadata -- the embedded information about how and when a file was created.

EXIF data in images. Most digital cameras and smartphones embed EXIF data in photos, including the device model, date, GPS coordinates, and camera settings. AI-generated images typically lack this data entirely, or contain generic metadata that does not correspond to any real device. Tools like Jeffrey's EXIF Viewer or ExifTool can extract and display this information.

Video container metadata. Video files contain metadata about the encoding software, resolution, frame rate, and creation date. A video that claims to be raw footage from a phone but was encoded with professional editing software warrants scrutiny. Inconsistencies between the claimed provenance of a video and its technical metadata are significant red flags.

Compression artifacts. Each time a video or image is saved, it undergoes compression that leaves characteristic artifacts. A file that has been processed through a deepfake pipeline and then re-encoded will often show compression patterns inconsistent with a file that was captured directly from a camera. Multiple generations of compression produce distinctive layered artifacts visible under close examination.

C2PA and Content Credentials. The Coalition for Content Provenance and Authenticity (C2PA) standard, adopted by major camera manufacturers and platforms, embeds cryptographic provenance information directly into media files. Content Credentials allow you to verify the origin and edit history of an image or video. While not yet universal, checking for C2PA credentials is becoming an increasingly valuable first step in verification.

Free Detection Tools You Can Use Today

Several tools are available to help non-experts evaluate whether media has been synthetically generated or manipulated.

  1. Microsoft Video Authenticator. Developed in partnership with the AI Foundation, this tool analyzes photos and videos and provides a confidence score indicating the likelihood that the media has been artificially manipulated. It works by detecting the blending boundary and subtle fading or greyscale elements that the human eye may not catch.

  2. Deepware Scanner. A free online tool that allows you to paste a video URL or upload a file for deepfake analysis. It uses multiple detection models to evaluate the video and provides a probability assessment. It is particularly useful for quickly scanning social media videos.

  3. InVID Verification Plugin. Originally developed for journalists, InVID is a browser extension that provides a suite of tools for verifying video and image content. It offers reverse image search across multiple engines, video fragmentation for frame-by-frame analysis, metadata extraction, and magnification tools. It is available for Chrome and Firefox and is free to use.

  4. FotoForensics. This web-based tool performs Error Level Analysis (ELA), which highlights areas of an image that have been modified at different compression levels. Regions that have been edited or pasted in will often show different error levels than the rest of the image.

  5. Hive Moderation. Hive offers a free AI-generated content detection tool that can assess whether text, images, or video were produced by AI. It provides probability scores for various generative models.

  6. Sensity AI. Focused specifically on deepfake detection, Sensity provides tools for analyzing both face-swap deepfakes and entirely synthetic faces. Its platform is used by news organizations and government agencies.

It is important to approach these tools with appropriate expectations. No detection tool is infallible, and the best deepfakes will evade automated detection. These tools are most useful as one component of a broader verification workflow rather than as standalone arbiters of truth.

Reverse Image Search: Tracing the Origins

Reverse image search remains one of the most effective and accessible techniques for verifying visual media. By uploading an image to a search engine, you can find where else it has appeared online, which can reveal whether an image has been recycled from an unrelated context or has a documented history predating the current claim.

When conducting a reverse image search, pay attention to the dates of the earliest results. If a supposedly breaking news photo appears in results from months or years earlier, it has been recycled. If the image appears in no results at all, it may be newly generated -- either legitimately or synthetically.

For video, you can extract individual frames using tools like InVID and then run those frames through reverse image search. This technique is particularly effective for identifying footage that has been repurposed from a different event or location.

A Verification Workflow for Journalists and Researchers

When you encounter a piece of media that you need to verify -- whether for publication, for sharing, or simply for your own confidence -- a structured workflow produces more reliable results than ad hoc checking.

  1. Assess the source. Where did the media first appear? Who shared it, and what is their track record? Is the original source identifiable?
  2. Check the metadata. Extract and examine the file's metadata for device information, timestamps, and provenance credentials.
  3. Conduct a reverse image or video search. Determine whether the media has appeared elsewhere in a different context.
  4. Examine the content closely. Apply the visual and audio analysis techniques described above. Look at the eyes, edges, hair, teeth, skin, and background. Listen for prosody, breathing, and phoneme issues.
  5. Run detection tools. Submit the media to one or more automated detection tools and note their assessments. Do not rely on a single tool.
  6. Seek corroboration. Can the event depicted in the media be confirmed by independent sources? Are there other angles, eyewitness accounts, or contemporaneous records that support or contradict it?
  7. Consult experts when stakes are high. For media that could significantly impact public discourse, elections, or individual reputations, consult with digital forensics specialists or organizations like the Digital Forensic Research Lab or Bellingcat.

This workflow does not guarantee that you will catch every deepfake, but it dramatically reduces the probability of being deceived. The key principle is layered verification -- no single check is sufficient, but multiple independent checks create a robust assessment.

The Arms Race: Why Detection Will Always Be a Moving Target

It is critical to understand that deepfake detection is fundamentally an adversarial problem. Every detection technique described in this guide exploits a current weakness of generative models, and the developers of those models are actively working to eliminate those weaknesses. The artifacts visible in today's deepfakes will be less visible in next year's deepfakes. Detection tools that perform well today may be evaded by tomorrow's generation models.

This does not mean detection is futile. It means that detection must be understood as an ongoing practice rather than a solved problem. The tools, techniques, and visual tells described here are current as of 2026, and they will need to be updated as the technology evolves.

Several trends are shaping the future of this arms race:

The most resilient defense is not any single tool or technique but the habit of critical evaluation -- the reflex to question, verify, and seek corroboration before accepting media at face value.

Building Your Deepfake Detection Skills

Like any skill, detecting deepfakes improves with practice. Several resources can help you develop your eye and ear for synthetic media.

The goal is not to become suspicious of everything you see and hear, but to develop a calibrated sense of when additional scrutiny is warranted. In most cases, the media you encounter is authentic. But in the cases where it is not, having the skills and tools to identify the deception is increasingly a matter of civic necessity.

For a comprehensive exploration of deepfakes, media manipulation, and the broader landscape of digital literacy, read our free Media Literacy textbook. It provides structured, in-depth coverage of the techniques, technologies, and critical thinking frameworks you need to navigate the modern information environment with confidence.