Deepfake Detection Project Guidance
Deepfake Detection Project Guidance
Executive Summary
The landscape of digital media authenticity has been profoundly reshaped by the
rapid advancements in deepfake technology. These synthetic media, generated
through sophisticated artificial intelligence algorithms, pose significant challenges to
digital security and information integrity. The core objective of deepfake detection is
to reliably distinguish between genuine and manipulated content, a task complicated
by the continuous evolution of forgery techniques.
Further sophistication led to the integration of synthetic voices with visual content,
resulting in highly convincing audio-visual deepfakes. Concurrently, neural textures
emerged, utilizing deferred neural rendering to push the boundaries of deepfake
realism.1 However, challenges persisted, including issues like mismatched lip
movements and visible splicing boundaries, particularly under complex lighting
conditions.1
The continuous progression from autoencoders to GANs and then to diffusion models
illustrates a dynamic often described as an "arms race" between deepfake creators
and their detectors.2 Early generative methods produced more overt, "low-order"
artifacts that were relatively straightforward to identify.1 As generative models became
more sophisticated, the nature of these detectable forgery traces evolved, shifting
from easily discernible flaws to increasingly subtle, high-frequency, or temporal
anomalies. For instance, diffusion models are noted for generating fewer detectable
artifacts and lacking the distinct "grid-like frequency artifacts" commonly associated
with GANs.3 This ongoing evolution means that any effective deepfake detection
system cannot rely on static, method-specific artifact identification. Instead, it must
be designed with inherent adaptability, perhaps through modular architectures or
continuous learning pipelines, focusing on fundamental, invariant traces of the
generative process rather than superficial cues. The core challenge is not merely
detecting current deepfakes but anticipating and countering future ones.
For a machine learning project focused on deepfake image detection and model
generalization, several publicly available datasets are highly recommended:
● FaceForensics++ (FF++): This is a foundational forensics dataset comprising
1000 original video sequences sourced from YouTube. These videos have been
manipulated using four automated face manipulation methods: Deepfakes,
Face2Face, FaceSwap, and NeuralTextures.6 FF++ also includes various
compression levels (raw/0, c23, c40), which are crucial for evaluating model
robustness to real-world data degradation.25 The raw video data is approximately
38.5GB, and raw extracted images can be as large as 2TB. However, compressed
versions (c23: ~10GB, c40: ~2GB) are significantly smaller and more
manageable.25 FF++ is a widely used benchmark, and models trained on it
(especially in combination with DFDC) have demonstrated comparably better
generalization capabilities.18 For Google Colab compatibility, using the
compressed versions is highly recommended due to storage and processing
limitations. Efficient loading strategies, such as
ImageDataGenerator or mounting Google Drive, are essential.19
● Deepfake Detection Challenge (DFDC) Dataset: This is a large-scale dataset
designed specifically for deepfake detection, containing over 100,000 videos.6 It
offers a preview version with 5,000 videos and two facial modification algorithms,
and a full version with 124,000 videos featuring eight different facial modification
algorithms.22 The videos are sourced from paid actors, emphasizing data
diversity.22 Alongside FF++, DFDC is recognized for equipping models with
superior generalization capabilities.18 Given its large size, similar considerations as
FF++ apply for Google Colab, including the need for efficient loading and
potentially a paid Colab plan for full dataset processing.19
● Deep Fake Detection (DFD) Entire Original Dataset: This dataset provides a
comprehensive collection of video sequences for deepfake detection,
downloaded from the official FaceForensics server. Its primary purpose is to
address the gap in available video-based original datasets for deepfake
detection.29 While the exact size is not specified in the available information, it is
implied to be of "substantial volume".29 It is suitable for training and evaluating
deep learning models for identifying manipulated media.29 As a video dataset, it
will likely require similar storage and efficient loading strategies as FF++ and
DFDC.19
● DFWild-Cup Competition Dataset: This dataset was provided for the 2025
Signal Processing Cup and contains a diverse range of deepfake images
generated from multiple deepfake image generators.20 Its explicit purpose is to
emphasize generalization in deepfake detection research.20 While its exact size is
not detailed, its design for a competition implies a substantial and challenging
collection of data. It is particularly valuable for projects focused on robust
generalization.
● StyleGAN-StyleGAN2 Deepfake Face Images Dataset: This dataset combines
and augments two existing datasets to ensure diversity, balance, and
generalizability.16 It includes 140,000 images (70,000 real from Nvidia's Flickr
dataset, 70,000 fake generated by StyleGAN and sourced from the DeepFake
Detection Challenge Discussion on Kaggle) and an additional 1,288 images (700
real from Unsplash, 588 fake from thispersondoesnotexist.com, utilizing
StyleGAN2).16 The dataset is further augmented with techniques like rotation,
shifting, brightness modification, zooming, cropping, and flipping, resulting in a
final size of 12,890 images.16 This dataset is explicitly curated for training and
evaluating deepfake detection models, studying GAN-generated images, and
exploring generalization techniques across multiple sources.16 Its relatively smaller
size makes it highly compatible with Google Colab for direct use or Google Drive
mounting.19
● Diffusion-generated Deepfake Detection dataset (D3): This large-scale
dataset addresses limitations in diversity and quantity found in older datasets. It
contains nearly 2.3 million records and 11.5 million images.21 Each record includes
a prompt, a real image, and four images generated by state-of-the-art
open-source diffusion models, specifically Stable Diffusion 1.4, 2.1, XL, and
DeepFloyd IF.21 The dataset also incorporates varying aspect ratios and
compression methods (BMP, GIF, JPEG, TIFF, PNG) to mimic real-world
distributions.21 D3 is crucial for addressing the detection of deepfakes generated
by diffusion models, which often lack the grid-like artifacts common in GANs.3
Given its very large size, careful memory management, potentially using
DiskDataset or iterating over samples, and mounting Google Drive for storage are
essential for Google Colab users.27
● DeepFakeFace (DFF) Dataset: This is a meticulously curated collection of
artificial celebrity faces generated using cutting-edge diffusion models, including
Stable Diffusion Inpainting, InsightFace, and Stable Diffusion V1.5.31 It also
includes original real images from the IMDB-WIKI dataset.31 The dataset is
distributed across four zip files, each containing 30,000 images, totaling 120,000
images.31 DFF is specifically designed to address challenges posed by deepfakes
generated by diffusion models.31 Its manageable size makes it suitable for Google
Colab, especially if mounted from Google Drive.27
● DF40: A Next-Generation Comprehensive Dataset: This dataset is a significant
advancement, comprising 40 distinct deepfake techniques, including recently
released State-of-the-Art (SOTA) methods.32 It is categorized into Face-swapping
(10 methods), Face-reenactment (13 methods), Entire Face Synthesis (12
methods, including StyleGAN2, StyleGAN3, StyleGAN-XL, SD-2.1, DDPM,
MidJourney6), and Face Edit (5 methods).32 DF40 provides both processed data
and original fake videos.32 Accepted by NeurIPS 2024, it is explicitly designed as a
"next-generation" dataset for comprehensive deepfake detection, making its
breadth of techniques exceptionally valuable for robust generalization research.32
Given its comprehensive nature, DF40 will likely be very large, necessitating
significant storage (Google Drive) and efficient processing strategies for Google
Colab users.32
The shift towards datasets that explicitly include deepfakes generated by diffusion
models is imperative. While foundational datasets like FF++ and DFDC are crucial,
newer datasets such as D3, DFF, and particularly DF40 directly address the evolving
threat. Diffusion models produce high-quality and diverse outputs that often do not
exhibit the noticeable grid-like artifacts in the frequency domain that older GANs do.3
This signifies a critical evolution in deepfake generation, meaning that detection
models trained solely on older GAN-based datasets will inherently struggle to
generalize to these newer, more sophisticated fakes. DF40, with its inclusion of 40
distinct techniques, including recent SOTAs from diffusion models, serves as an
excellent benchmark for training and evaluating next-generation deepfake detectors
that can handle the evolving threat landscape.32
DiskDataset for data larger than memory 30 are essential for efficient memory and I/O
management. Selecting datasets must involve a trade-off between diversity/size and
Colab's practical constraints. Adopting efficient data handling practices is necessary
to avoid runtime crashes and ensure a smooth development workflow. For very
large-scale experiments, a paid Colab plan or alternative computing resources might
eventually be necessary.
diversity that mimics the variations found in real-world deepfakes and their
dissemination, such as different lighting, angles, and compression artifacts. By
applying these transformations, the model is compelled to learn features that are
invariant to these common variations, directly improving its generalization to unseen,
real-world deepfakes. This is particularly crucial when obtaining truly diverse
"in-the-wild" data is challenging or impossible. Image augmentations have been
shown to be particularly helpful in improving the performance of Transformer models,
which are increasingly recognized for their strong generalization capabilities in
deepfake detection.18 Strategic and diverse data augmentation leads to models
learning more robust and invariant features, which in turn significantly improves their
generalization performance on out-of-distribution data.
Deepfake detection has traditionally relied on analyzing visual cues in the spatial
domain. However, the increasing sophistication of generative models necessitates a
complementary approach that can uncover subtle, often imperceptible, forgery
traces. Frequency domain analysis offers precisely this advantage.
While the spatial domain, which includes pixel values, colors, and textures, offers rich
feature representations for detecting image details and structures, most
spatial-based models tend to overfit to specific forgery characteristics.35 This
limitation often leads to poor generalization, especially when dealing with low-quality
or compressed images.35
Beyond merely detecting whether an image is fake, the ability to attribute a deepfake
to its generative source is becoming increasingly important for forensic analysis,
intellectual property protection, and countering misinformation.
The complexity and sheer number of available generative techniques, coupled with
the scarcity of high-quality, diverse open-source datasets specifically for this task,
make training and benchmarking synthetic image source attribution models highly
challenging.45 Current methods struggle with generalization, particularly in open-set
scenarios involving unknown or novel generators.45 Furthermore, deepfake detectors
are vulnerable to counter-forensic attacks, where malicious actors intentionally exploit
the limitations of existing detection methods.45
Various methodologies are being developed to identify the training dataset or specific
generative model used to create deepfake images. These approaches often leverage
subtle, inherent "fingerprints" left by the generative process.
One novel forensic framework for identifying the training dataset (e.g., CelebA or
FFHQ) of GAN-generated images employs interpretable feature analysis.44 This
pipeline integrates spectral transforms, color distribution metrics, and local feature
descriptors to extract discriminative statistical signatures.44
● Spectral Transforms: Discrete Cosine Transform (DCT) and Fast Fourier
Transform (FFT) are applied to convert images to their spectral representations.
This allows for the identification of dataset-specific periodic artifacts and
compression-related signatures. For example, real datasets show smoother
frequency decay, while GAN-generated images often display localized
high-frequency artifacts or cross-shaped spectral artifacts, likely due to
convolutional upsampling patterns.38
● Color Distribution Metrics: Normalized RGB histograms capture color
information, revealing that real datasets maintain characteristic channel skews,
whereas GAN-generated images tend to smooth and normalize color profiles,
indicating implicit regularization strategies during GAN training.44
● Local Feature Descriptors: Scale-Invariant Feature Transform (SIFT) descriptors
are employed to model textural and structural properties.44
These extracted features are then concatenated into a single feature vector and
standardized before being fed into supervised machine learning classifiers such
as Random Forest, Support Vector Machines (SVM), K-Nearest Neighbors (K-NN),
and XGBoost.38 This framework has achieved high accuracy (98-99%) in both
binary classification (real vs. synthetic) and multi-class dataset attribution across
diverse GAN architectures.44 Frequency-domain features (DCT/FFT) are
particularly dominant in capturing dataset-specific artifacts.38
By focusing on these globally consistent features, DNA-Det can isolate the unique
"fingerprints" left by the GAN's underlying architecture, leading to robust attribution
even in challenging cross-test setups (cross-seed, cross-loss, cross-finetune, and
cross-dataset).46
For diffusion models, which are increasingly prevalent, specific attribution challenges
arise because they often do not exhibit the same grid-like frequency artifacts as
GANs.3 However, research is progressing on identifying and attributing images from
text-to-image synthesis models like Stable Diffusion and DeepFloyd IF.48 Methods like
"DeepGuard" propose multi-classification techniques to attribute fake images to their
source models, achieving high accuracy.48 This involves training models to distinguish
between synthetic images created by various text-to-image models (e.g., Stable
Diffusion 1.5, SDXL, DeepFloyd IF) and authentic ones.48 The open-source nature of
models like Stable Diffusion allows for community involvement in developing detection
methods, focusing on inconsistencies like distorted fingers/hands, unnatural limbs,
blurry backgrounds, or garbled text, which are common artifacts.49
5. Practical Considerations for Google Colab Environment
For deepfake detection projects, which often involve substantial datasets, efficient
data handling is paramount. Datasets like FaceForensics++ can be extremely large
(e.g., ~2TB for raw extracted images, ~500GB for raw videos).25 Directly loading such
volumes into Colab's ephemeral runtime memory is often impractical.
The recommended strategy for handling large datasets in Colab is to load data
directly from Google Drive by using the mount drive method.23 This imports all
data from the user's Drive to the Colab runtime instance, providing persistent storage
beyond the session's ephemeral nature.23 For video datasets like FF++ and DFDC, it is
highly advisable to use
compressed versions (e.g., FF++ c23 or c40, which are significantly smaller at ~10GB
and ~2GB respectively) to reduce storage and processing demands.25
When working with image data, the ImageDataGenerator from Keras/TensorFlow can
be invaluable. Its flow_from_directory method helps read images from directories and
transform them into normalized NumPy arrays on the fly, reducing the need to load
the entire dataset into memory at once.19 This method allows specifying batch size
and target image size, which can be adjusted based on model requirements.19 For
datasets larger than memory, DeepChem's
DiskDataset class provides tools for efficiently working with data saved to disk,
ensuring it can be accessed even if the total amount exceeds available memory.30 This
approach involves iterating over the dataset, loading only small batches at a time,
processing them, and then freeing memory before loading the next batch.30
Key Conclusions:
● Dynamic Threat Landscape: Deepfake technology is rapidly advancing, with
diffusion models introducing new types of artifacts that differ from those
generated by older GANs. This means that detection models must constantly
adapt, and reliance on methods that overfit to "toxic" low-order interactions or
specific, transient artifacts will lead to poor generalization.
● Dataset Diversity is Paramount: The sheer volume of data is not enough.
Generalizable models require training on diverse datasets that encompass a wide
range of deepfake generation techniques (including diffusion models), varying
compression levels, and "in-the-wild" characteristics. Datasets like FF++, DFDC,
D3, DFF, and particularly DF40 are crucial for this purpose.
● Frequency Domain Analysis is Essential: Spatial-domain analysis alone is
insufficient. Frequency domain analysis offers a complementary lens to uncover
subtle, often imperceptible, forgery traces and enhance robustness against image
compression. The development of frequency-aware architectures and
spatial-frequency fusion techniques is critical.
● Source Attribution is a Growing Necessity: Moving beyond binary
classification, the ability to identify the specific generative model or training
dataset of a deepfake is vital for forensic analysis, accountability, and intellectual
property protection. This requires specialized attribution methodologies that can
detect architectural fingerprints.
● Practicality in Cloud Environments: For researchers utilizing platforms like
Google Colab, efficient data handling strategies (e.g., Google Drive mounting,
ImageDataGenerator, DiskDataset) are not merely optimizations but necessities
for managing large datasets and ensuring productive research.
Actionable Recommendations:
1. Prioritize Diverse Dataset Acquisition and Generation:
○ Actively seek and integrate datasets that include deepfakes generated by
diffusion models (e.g., D3, DFF, DF40) to prepare models for the latest
generation of forgeries.
○ Emphasize datasets with "in-the-wild" characteristics, including varying
resolutions, compression levels, and diverse environmental factors, to
enhance real-world applicability.
○ Consider synthetic data generation and strategic data augmentation
(e.g., rotation, shifting, brightness, cropping, flipping) to further expand
dataset diversity and simulate unseen variations, especially for training
Transformer models.
2. Adopt Spatial-Frequency Fusion Architectures:
○ Implement and experiment with hybrid detection frameworks that integrate
both spatial and frequency domain analysis (e.g., SFDT, SFCL, FMSI, CFNSFD).
This dual-domain approach is crucial for capturing both overt visual
inconsistencies and subtle spectral artifacts.
○ Explore frequency-aware learning techniques like FreqNet, which compel
models to learn generalized, source-agnostic features in the frequency
domain, rather than overfitting to specific GAN-induced grid patterns.
3. Implement Generalization-Enhancing Training Strategies:
○ Utilize ensemble learning approaches by combining predictions from
multiple asymmetric models. This strategy provides more stable and reliable
performance across diverse, unseen datasets, mitigating the risk of individual
model failures.
○ Investigate meta-learning and domain adaptation techniques (e.g.,
OWG-DS, MCW) to enable models to learn from limited labeled data and
adapt effectively to large-scale unlabeled target domains, addressing the
critical domain shift problem.
○ Leverage self-supervised pre-training strategies (e.g., DINO) for backbone
models to learn superior feature representations that generalize better across
different deepfake types.
4. Develop and Integrate Deepfake Attribution Capabilities:
○ Incorporate methodologies for source dataset attribution (e.g., using
spectral transforms, color distribution metrics, and local feature descriptors)
to identify the training data used for GAN-generated images.
○ Explore techniques for generative model architecture attribution (e.g.,
DNA-Det) to pinpoint the specific deepfake generation tool, even if it has
been finetuned or retrained. This is crucial for forensic analysis and
accountability.
5. Optimize for Google Colab Environment:
○ Always mount Google Drive for persistent storage of large datasets and
model checkpoints.
○ Prioritize using compressed versions of large video datasets (e.g., FF++
c23/c40) to manage resource consumption.
○ Employ efficient data loading mechanisms like ImageDataGenerator for
image data and consider DiskDataset for datasets exceeding memory
capacity.
○ Utilize callbacks (e.g., EarlyStopping) during training to prevent overfitting
and optimize computational resource usage.
○ Be prepared to consider Colab Pro/Pro+ for more demanding experiments
that require guaranteed resources and longer runtimes.
Works cited
1. Generative Artificial Intelligence and the Evolving Challenge of ..., accessed July
19, 2025, https://www.mdpi.com/2224-2708/14/1/17
2. The Evolution of Deepfake Technology: From Simple Face Swaps to AI-Powered
Illusions | by SingularityNET Ambassadors - Medium, accessed July 19, 2025,
https://medium.com/singularitynet-ambassador-community-page/the-evolution-
of-deepfake-technology-from-simple-face-swaps-to-ai-powered-illusions-59fe
88d89b96
3. Diffusion Deepfake - OpenReview, accessed July 19, 2025,
https://openreview.net/pdf/add2d2cf12b751ad53cb1501fa1937a9b42db92b.pdf
4. Towards Understanding the Generalization of ... - CVF Open Access, accessed
July 19, 2025,
https://openaccess.thecvf.com/content/ICCV2023/papers/Yao_Towards_Understa
nding_the_Generalization_of_Deepfake_Detectors_from_a_Game-Theoretical_IC
CV_2023_paper.pdf
5. Towards the Detection of Diffusion Model Deepfakes - SciTePress, accessed July
19, 2025, https://www.scitepress.org/Papers/2024/124220/124220.pdf
6. A SURVEY ON: DEEPFAKE DETECTION SYSTEM - IRJMETS, accessed July 19, 2025,
https://www.irjmets.com/uploadedfiles/paper//issue_11_november_2024/63699/fin
al/fin_irjmets1731397529.pdf
7. Towards Open-world Generalized Deepfake Detection: General Feature
Extraction via Unsupervised Domain Adaptation - arXiv, accessed July 19, 2025,
https://arxiv.org/html/2505.12339v1
8. www.jas.shu.edu.cn, accessed July 19, 2025,
https://www.jas.shu.edu.cn/EN/10.3969/j.issn.0255-8297.2025.03.007
9. Ensemble-Based Deepfake Detection using State-of-the-Art ... - arXiv, accessed
July 19, 2025, https://arxiv.org/abs/2507.05996
10.Comprehensive Evaluation of Deepfake Detection Models: Accuracy ..., accessed
July 19, 2025, https://www.mdpi.com/2076-3417/15/3/1225
11. A Comprehensive Survey of DeepFake Generation and Detection ..., accessed
July 19, 2025, https://www.icck.org/article/abs/jiap.2025.431672
12.A Comprehensive Evaluation of Deepfake Detection Methods: Approaches,
Challenges and Future Prospects | ITM Web of Conferences, accessed July 19,
2025,
https://www.itm-conferences.org/articles/itmconf/abs/2025/04/itmconf_iwadi202
4_03002/itmconf_iwadi2024_03002.html
13.Deepfakes Generation and Detection: A Short Survey - PMC, accessed July 19,
2025, https://pmc.ncbi.nlm.nih.gov/articles/PMC9863015/
14.A Survey: Deepfake and Current Technologies for ... - CEUR-WS.org, accessed
July 19, 2025, https://ceur-ws.org/Vol-3900/Paper9.pdf
15.Adaptive DCGAN and VGG16-based transfer learning for effective deepfake
detection - IET Digital Library, accessed July 19, 2025,
https://digital-library.theiet.org/doi/pdf/10.1049/icp.2025.1711?download=true
16.StyleGan-StyleGan2 Deepfake Face Images - Kaggle, accessed July 19, 2025,
https://www.kaggle.com/datasets/kshitizbhargava/deepfake-face-images
17.MCW: A Generalizable Deepfake Detection Method for Few-Shot ..., accessed
July 19, 2025, https://pmc.ncbi.nlm.nih.gov/articles/PMC10649340/
18.(PDF) Deepfake Detection: Analysing Model Generalisation Across ..., accessed
July 19, 2025,
https://www.researchgate.net/publication/376975669_Deepfake_Detection_Analy
sing_Model_Generalisation_Across_Architectures_Datasets_and_Pre-Training_Pa
radigms
19.DeepFake Detection.ipynb - Colab, accessed July 19, 2025,
https://colab.research.google.com/drive/1GRyxgtxpxGNLQkvcUcpF3cXEnFQSNpI
3?usp=sharing
20.Hybrid Deepfake Image Detection: A Comprehensive Dataset-Driven Approach
Integrating Convolutional and Attention Mechanisms with Frequency Domain
Features - arXiv, accessed July 19, 2025, https://arxiv.org/html/2502.10682v1
21.Diffusion-generated Deepfake Detection dataset (D 3 ) - AImageLab - Unimore,
accessed July 19, 2025,
https://aimagelab.ing.unimore.it/imagelab/go.asp?string=d3-dataset
22.[1910.08854] The Deepfake Detection Challenge (DFDC) Preview Dataset - ar5iv -
arXiv, accessed July 19, 2025, https://ar5iv.labs.arxiv.org/html/1910.08854
23.Deepfakes with the First Order Model Method - Colab - Google, accessed July
19, 2025,
https://colab.research.google.com/github/JaumeClave/deepfakes_first_order_mo
del/blob/master/first_order_model_deepfakes.ipynb
24.Frequency-Aware Deepfake Detection: Improving Generalizability through
Frequency Space Learning - arXiv, accessed July 19, 2025,
https://arxiv.org/html/2403.07240v1
25.FaceForensics/dataset/README.md at master - GitHub, accessed July 19, 2025,
https://github.com/ondyari/FaceForensics/blob/master/dataset/README.md
26.FaceForensics++ Dataset - Papers With Code, accessed July 19, 2025,
https://paperswithcode.com/dataset/faceforensics-1
27.neptune.ai, accessed July 19, 2025,
https://neptune.ai/blog/how-to-use-google-colab-for-deep-learning-complete-t
utorial#:~:text=To%20train%20complex%20models%2C%20you,where%20the%
20dataset%20is%20stored.
28.DFDC Dataset - Papers With Code, accessed July 19, 2025,
https://paperswithcode.com/dataset/dfdc
29.Deep Fake Detection (DFD) Entire Original Dataset - Kaggle, accessed July 19,
2025,
https://www.kaggle.com/datasets/sanikatiwarekar/deep-fake-detection-dfd-entir
e-original-dataset
30.Working With Datasets - Colab, accessed July 19, 2025,
https://colab.research.google.com/github/deepchem/deepchem/blob/master/exa
mples/tutorials/Working_With_Datasets.ipynb
31.OpenRL/DeepFakeFace · Datasets at Hugging Face, accessed July 19, 2025,
https://huggingface.co/datasets/OpenRL/DeepFakeFace
32.YZY-stack/DF40: Official repository for the next-generation ... - GitHub, accessed
July 19, 2025, https://github.com/YZY-stack/DF40
33.In-the-Wild Dataset - Deepfake Total, accessed July 19, 2025,
https://deepfake-total.com/in_the_wild
34.Google Colab, accessed July 19, 2025,
https://research.google.com/colaboratory/faq.html
35.Frequency-Domain Masking and Spatial Interaction for Generalizable Deepfake
Detection, accessed July 19, 2025, https://www.mdpi.com/2079-9292/14/7/1302
36.(PDF) Deepfake Detection Technology Integrating Spatial Domain and Frequency
Domain, accessed July 19, 2025,
https://www.researchgate.net/publication/390254350_Deepfake_Detection_Tech
nology_Integrating_Spatial_Domain_and_Frequency_Domain
37.Towards Generalizable Deepfake Detection with Spatial-Frequency Collaborative
Learning and Hierarchical Cross-Modal Fusion - arXiv, accessed July 19, 2025,
https://arxiv.org/html/2504.17223v1
38.Deepfake Forensic Analysis: Source Dataset Attribution and Legal Implications of
Synthetic Media Manipulation - arXiv, accessed July 19, 2025,
https://www.arxiv.org/pdf/2505.11110
39.[Papierüberprüfung] Frequency-Aware Deepfake Detection: Improving
Generalizability through Frequency Space Learning - Moonlight | AI Colleague for
Research Papers, accessed July 19, 2025,
https://www.themoonlight.io/de/review/frequency-aware-deepfake-detection-im
proving-generalizability-through-frequency-space-learning
40.[Literature Review] Frequency-Aware Deepfake Detection: Improving
Generalizability through Frequency Space Learning - Moonlight, accessed July 19,
2025,
https://www.themoonlight.io/review/frequency-aware-deepfake-detection-impro
ving-generalizability-through-frequency-space-learning
41.Deepfake Detection Technology Integrating Spatial Domain and Frequency
Domain | Frontiers in Computing and Intelligent Systems - Darcy & Roy Press,
accessed July 19, 2025, https://drpress.org/ojs/index.php/fcis/article/view/30019
42.(PDF) Deepfake Detection Based on the Adaptive Fusion of Spatial ..., accessed
July 19, 2025,
https://www.researchgate.net/publication/385649528_Deepfake_Detection_Base
d_on_the_Adaptive_Fusion_of_Spatial-Frequency_Features
43.arxiv.org, accessed July 19, 2025, https://arxiv.org/pdf/2403.07240v1
44.(PDF) Deepfake Forensic Analysis: Source Dataset Attribution and ..., accessed
July 19, 2025,
https://www.researchgate.net/publication/391856499_Deepfake_Forensic_Analysi
s_Source_Dataset_Attribution_and_Legal_Implications_of_Synthetic_Media_Mani
pulation
45.Deepfake attribution: On the source identification of artificially ..., accessed July
19, 2025,
https://www.researchgate.net/publication/356770322_Deepfake_attribution_On_t
he_source_identification_of_artificially_generated_images
46.Deepfake Network Architecture Attribution - Association for the ..., accessed July
19, 2025, https://cdn.aaai.org/ojs/20391/20391-13-24404-1-2-20220628.pdf
47.Deepfake Network Architecture Attribution | Proceedings of the AAAI
Conference on Artificial Intelligence, accessed July 19, 2025,
https://ojs.aaai.org/index.php/AAAI/article/view/20391
48.DeepGuard: Identification and Attribution of AI-Generated Synthetic ..., accessed
July 19, 2025,
https://researchportal.port.ac.uk/files/100827541/electronics-14-00665.pdf
49.Stable Diffusion Deepfakes: Creation and Detection | by Tahir - Medium, accessed
July 19, 2025,
https://medium.com/@tahirbalarabe2/stable-diffusion-deepfakes-creation-and-d
etection-15103f99f55d
50.deep-floyd/IF - GitHub, accessed July 19, 2025, https://github.com/deep-floyd/IF
51.Stability AI releases DeepFloyd IF, a powerful text-to-image model that can
smartly integrate text into images, accessed July 19, 2025,
https://stability.ai/news/deepfloyd-if-text-to-image-model