Blockchain + AI in Combating Deepfake Content Circulation
DOI:
https://doi.org/10.63345/sjaibt.v1.i2.102Keywords:
Blockchain, Deepfakes, Content Authenticity, C2PA, Content Credentials, Verifiable Credentials, DIDs, Zero-Knowledge Proofs, IPFS, AI DetectionAbstract
The rapid proliferation of AI-generated “deepfake” images, audio, and video is eroding public trust in digital media and amplifying risks to elections, markets, journalism, and personal safety. While AI detection models have improved, they face an adversarial “cat-and-mouse” problem and often struggle to generalize across manipulation methods and compression regimes. This manuscript proposes and analyzes a hybrid, end-to-end approach that couples upstream provenance and authenticity signals—anchored via open standards (e.g., C2PA Content Credentials) and decentralized ledgers—with downstream AI detection and moderation. The pipeline captures and signs media at source; binds verifiable, tamper-evident metadata; anchors cryptographic hashes on a public or consortium blockchain; stores originals off-chain with content addressing (e.g., IPFS/Filecoin); and fuses these trust signals with model-based detectors and policy engines at distribution edges. We situate the proposal within current regulation (e.g., EU AI Act transparency duties) and state-of-the-art methods (e.g., watermarking such as SynthID, Stable Signature, and Tree-Ring; deepfake detectors trained on DFDC and FaceForensics++), highlighting both strengths and known attack vectors against watermarking that motivate layered defenses. A simulation-based evaluation illustrates that combining provenance signals with video-level transformer detectors can raise F1 from 0.85 to 0.92 while cutting false positives by ~41% in a balanced test set, primarily by rejecting credential-mismatched or hash-divergent media before expensive model inference. We further discuss privacy-preserving verification using W3C Verifiable Credentials (VC 2.0), Decentralized Identifiers (DIDs), and selective-disclosure with zero-knowledge proofs. The findings make a practical case for “trust by design” built on open standards, decentralized integrity proofs, and robust AI detection, implemented as a policy-aware defense-in-depth stack for platforms and newsrooms.
Downloads
References
• Adobe. (2025, July 15). Content Credentials overview. https://helpx.adobe.com/creative-cloud/apps/adobe-content-authenticity/content-credentials/overview.html
• Adobe. (2024, October 8). Introducing Adobe Content Authenticity: A free web app…. https://blog.adobe.com
• Coalition for Content Provenance and Authenticity (C2PA). (2025, May 1). C2PA Technical Specification v2.2. https://spec.c2pa.org
• Coalition for Content Provenance and Authenticity (C2PA). (n.d.). C2PA—Verifying media content sources. https://c2pa.org
• Content Authenticity Initiative (CAI). (n.d.). How it works. https://contentauthenticity.org/how-it-works
• DeepMind. (n.d.). SynthID. https://deepmind.google/science/synthid/
• Dolhansky, B., et al. (2020). The DeepFake Detection Challenge (DFDC) dataset. arXiv:2006.07397.
• European Parliament. (2025, February 19). EU AI Act: first regulation on artificial intelligence. https://www.europarl.europa.eu
• FaceForensics++. (2019). Learning to detect manipulated facial images (Rössler, A., et al.). arXiv:1901.08971.
• IEEE Spectrum. (2023). Detection stays one step ahead of deepfakes—for now. https://spectrum.ieee.org/deepfake
• Meta AI. (2020). DFDC dataset. https://ai.meta.com/datasets/dfdc/
• The Verge. (2024, Aug. 21). A system can sort real photos from AI fakes—why aren’t platforms using it? https://www.theverge.com
• The Verge. (2024, Oct. 23). Google open-sourced its watermarking tool for AI-generated text. https://www.theverge.com
• Truepic. (2023–2025). Technology for authenticity & C2PA conformance. https://www.truepic.com
• W3C. (2025, May 15). Verifiable Credentials Data Model 2.0. https://www.w3.org/TR/vc-data-model-2.0/
• W3C. (2022, July 19). Decentralized Identifiers (DID) v1.0. https://www.w3.org/TR/did-core/
• Wen, Y., Kirchenbauer, J., Geiping, J., & Goldstein, T. (2023). Tree-Ring Watermarks: Fingerprints for diffusion images that are invisible and robust. NeurIPS.
• Fernandez, P., Couairon, G., Jégou, H., Douze, M., & Furon, T. (2023). The Stable Signature: Rooting watermarks in latent diffusion models. ICCV.
• Wired. (2023). Researchers tested AI watermarks—and broke all of them. https://www.wired.com
• IPFS Docs. (n.d.). How IPFS works. https://docs.ipfs.tech/concepts/how-ipfs-works/
Downloads
Published
Issue
Section
License
Copyright (c) 2024 Scientific Journal of Artificial Intelligence and Blockchain Technologies

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.
The license allows re-users to share and adapt the work, as long as credit is given to the author and don't use it for commercial purposes.