Chapter 18: Key Takeaways — Deepfakes, Synthetic Media, and Emerging Threats

The Spectrum of Synthetic Media

1. "Deepfake" describes only one part of a broader spectrum of synthetic and manipulated media. The spectrum ranges from cheap fakes (low-technology manipulations like speed changes and selective editing) through shallowfakes (basic AI-assisted modifications) to deepfakes (AI-generated content replacing faces, voices, or bodies). Understanding this full spectrum is essential because cheap fakes — simple, fast, and cheap to produce — have often caused as much or more harm than sophisticated deepfakes, precisely because they can be produced and deployed more rapidly.

2. The 2017 Reddit deepfake transition was enabled by the convergence of available technology, accessible hardware, abundant training data, and open-source community development. The deepfake era began not with any single technical breakthrough but with the convergence of available GAN architecture (published 2014), consumer-grade GPU hardware, massive public image archives of celebrities and public figures, and an open-source community that rapidly improved and distributed accessible tools. This convergence transformed deepfake production from a specialist capability to a consumer technology.

3. The direction of travel is consistently toward higher quality, lower cost, and wider accessibility. Both generation quality and the data efficiency of generation systems are improving continuously. The technologies that are difficult now will not be difficult in three to five years. Policy frameworks that assume current quality limitations will be inadequate for the near-term threat landscape.


The Technology

4. Generative Adversarial Networks (GANs) work through adversarial training between a generator and discriminator. The generator produces synthetic content; the discriminator attempts to identify it as synthetic; the generator is trained to defeat the discriminator. This competitive dynamic produces increasingly realistic output but also characteristic limitations (mode collapse, training instability) and characteristic artifacts that forensic analysis can detect — for now.

5. Diffusion models have largely superseded GANs for high-quality image generation and are increasingly applied to video. Systems like Stable Diffusion, DALL-E, and Midjourney use diffusion-based architectures, which provide better training stability, higher quality, and text-conditioning capability. The development of text-to-video systems (Sora and successors) represents the next major capability expansion, enabling synthetic video generation from text prompts without target-specific training data.

6. Voice cloning is technically more mature and practically more widely deployed as a fraud tool than video deepfakes. Current voice cloning systems require very short audio samples, can be deployed at telephone scale, and exploit limited voice verification cultures in business and social contexts. The 2019 CEO voice fraud (€220,000) and the wave of grandparent scams using cloned children's voices demonstrate that audio deepfake fraud is not a future threat but a current operational harm.


The Harms

7. NCII deepfakes are the predominant form of deepfake harm by volume, targeting overwhelmingly women and girls. Approximately 96% of deepfakes on the public internet as of 2019 were non-consensual pornographic imagery targeting women. NCII deepfakes require only a publicly accessible photograph to produce — dramatically expanding the potential victim population compared to traditional revenge porn. The harms are severe and documented: PTSD, professional consequences, social withdrawal, and in some cases physical safety risks.

8. The Liar's Dividend is a second-order harm that may be more consequential than primary deepfake harm. Chesney and Citron's "Liar's Dividend" concept identifies how deepfake technology harms truth-telling even when fake videos are quickly debunked: the existence of convincing deepfakes provides a general-purpose denial mechanism for those who actually committed recorded misconduct. Every politician who wants to dismiss authentic damaging video now has a technically-sounding basis for claiming it is synthetic.

9. Political deepfakes have not yet produced documented large-scale electoral manipulation, but the threshold for real harm is declining. Documented political deepfakes — the Zelensky surrender video, the Gabon Bongo controversy — have either been crude and quickly debunked, or have had consequences mediated through the deepfake allegation rather than the deepfake itself. As quality improves and debunking becomes harder, the potential for genuine electoral manipulation grows.

10. Voice cloning fraud in financial contexts represents an already-operational harm at scale. Unlike speculative future harms from perfect deepfakes, voice cloning financial fraud is happening now, at scale, with documented billion-dollar aggregate losses. Business email compromise enhanced with voice cloning, grandparent scams, and CEO impersonation represent a mature fraud ecosystem that existing anti-fraud frameworks were not designed for.


Detection and Provenance

11. Artifact-based deepfake detection is subject to a continuous arms race with generation technology. Every specific artifact that detectors learn to identify can be used as a training signal to suppress that artifact in generators. Detection systems trained on one generation of deepfakes consistently fail to generalize to subsequent generations. This arms race dynamic means artifact-based detection will always lag behind generation capability.

12. C2PA content provenance addresses the authenticity problem at the creation level rather than through artifact analysis. By embedding cryptographically signed metadata at the point of content creation, C2PA enables positive authentication of content from trusted sources. The signatures are tamper-evident: altering the content invalidates the signature. C2PA proves authenticity for signed content; the absence of credentials does not prove inauthenticity. Full ecosystem adoption remains the key challenge.

13. No single detection approach is reliable across all contexts and generation systems. The most robust approach combines multiple methods: artifact analysis (as a preliminary screen), physiological signal detection (where applicable), content provenance verification (where credentials are available), source authentication (corroborating through the publishing entity's verified channels), and institutional fact-checking. No individual method is sufficient.

14. The detection arms race is structurally unfavorable to defenders. Attackers need only one approach that produces convincing synthetic media; defenders need to reliably detect all synthetic media. The asymmetry favors the attacker in most adversarial scenarios, which is why provenance-based approaches (which establish authenticity from trusted sources rather than detecting inauthenticity) are increasingly important.


15. U.S. NCII law is a patchwork of state legislation with significant gaps and enforcement challenges. Over 20 states have NCII laws, with varying scope, penalties, and civil cause of action provisions. Federal legislation has been proposed but not enacted as of early 2025. The fragmented landscape leaves many victims without adequate legal recourse, and criminal prosecution addresses only a tiny fraction of actual harm.

16. Platform policies on synthetic media are necessary but insufficient without robust enforcement capacity. Major platforms have NCII and deepfake policies, but enforcement faces fundamental challenges: the volume of uploaded content far exceeds human review capacity; automated detection has significant error rates; and specialized websites outside mainstream platforms are not addressed by platform policies.

17. The open-source AI ecosystem fundamentally limits platform-level and regulatory responses. Once generation models are publicly released as open-source, deployment on local hardware outside any platform's control is possible. No platform policy or regulatory mandate can prevent determined bad actors from using locally deployed open-source tools to produce deepfakes. Regulatory frameworks that depend on controlling access to the technology are structurally limited.

18. International coordination is insufficient for a harm that is inherently cross-border. NCII deepfake production, hosting, and distribution are frequently cross-border operations, exploiting jurisdictional arbitrage. Current international legal frameworks — including the Budapest Convention on Cybercrime — do not adequately address synthetic media. The absence of effective international coordination is a significant gap in the remediation landscape.


Future Trajectories and Epistemological Implications

19. The presumptive authenticity of audiovisual evidence is being fundamentally eroded. For most of recorded history, video and audio recordings carried an implicit claim to authenticity. That implicit claim is dissolving. This erosion has profound implications for legal systems (which rely on audiovisual evidence), journalism (which depends on documenting truth), democratic accountability (which requires documenting officials' actions), and everyday social trust.

20. Generalized skepticism is not the appropriate response to synthetic media. The correct response to the availability of convincing deepfakes is not to distrust all audiovisual content — this plays into the Liar's Dividend by giving genuine wrongdoers cover. The appropriate response is calibrated verification: asking for provenance credentials where available, corroborating through multiple independent sources, and relying on trusted institutional validators for high-stakes assessments.

21. Building verification cultures and institutions is more durable than purely technical solutions. Technical detection tools will always face the arms race problem; regulatory frameworks will always face the open-source gap. The most durable response involves building widespread understanding of how to verify information, developing institutions with expertise in media authentication, and creating verification cultures in journalism, law, and civic life.

22. The epistemological challenge of synthetic media creates opportunities for deliberate exploitation. The Liar's Dividend is not only used passively (people dismiss authentic footage because they know deepfakes exist) but is actively exploited by sophisticated actors who deploy the "deepfake accusation" as a political or legal strategy. Understanding this exploitation as a deliberate tactic — rather than as an inevitable byproduct of the technology — is essential for developing appropriate responses.

23. Deepfake harm is deeply gendered, and policy responses must account for this. The overwhelming preponderance of NCII deepfakes targets women. The chilling effects on women's public participation — in journalism, politics, and civic life — from the threat of deepfake-based harassment are documented. Policy frameworks that treat deepfake harm as gender-neutral will systematically underprotect the population most affected.

24. The appropriate policy response combines technical standards, regulatory frameworks, platform obligations, and media literacy investment. No single tool addresses the full landscape of synthetic media harm. Technical standards (C2PA) address provenance for compliant actors. Regulatory frameworks create deterrence and remediation paths. Platform obligations create immediate victim protection. And media literacy investment builds the population-level verification capacity that is the most durable protection against synthetic media disinformation.