Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

When Deepfakes Look Real: Detecting AI-Generated Faces with Unlabeled Data due to Annotation Challenges

Created by
  • Haebom

Author

Zhiqiang Yang, Renshuai Tao, Xiaolong Zheng, Guodong Yang, Chunjie Zhang

Outline

To address the labeled data dependency of existing deepfake detection methods, this paper proposes a Dual-Path Guidance Network (DPGNet) that leverages large-scale unlabeled data. DPGNet consists of two core modules: one that bridges the domain gap between face images generated by different generative models and the other that leverages unlabeled image samples. The first module, text-guided cross-domain alignment, integrates visual and textual embeddings into a domain-invariant feature space using learnable prompts. The second module, curriculum-driven pseudo label generation, dynamically leverages information-rich unlabeled samples. Furthermore, it prevents forgetting through cross-domain knowledge distillation. Experimental results on 11 datasets demonstrate that DPGNet outperforms the state-of-the-art (SoTA) method by 6.3%. This demonstrates the effectiveness of DPGNet in solving the increasingly realistic annotation problem of deepfake images by leveraging unlabeled data.

Takeaways, Limitations

Takeaways:
We present a novel method to effectively utilize large-scale, unlabeled data to improve deepfake detection performance.
We present effective strategies to reduce domain gaps and utilize unlabeled data (text-guided cross-domain alignment and curriculum-driven pseudo label generation).
Contributes to alleviating the difficulties of annotation work in the field of deepfake detection.
Achieves significant performance improvements over existing top-performing models.
Limitations:
Further verification of the generalization performance of the proposed method is needed.
Robustness evaluation of various types of deepfake generation models is needed.
Performance evaluation in real online social network environments is needed.
There may be a dependency on prompt engineering.
👍