The proliferation of synthetic facial imagery has intensified the need for robust Open-World DeepFake Attribution (OW-DFA), which aims to attribute both known and unknown forgeries using labeled data for known types and unlabeled data containing a mixture of known and novel types. However, existing OW-DFA methods face two critical limitations: 1) A confidence skew that leads to unreliable pseudo-labels for novel forgeries, resulting in biased training. 2) An unrealistic assumption that the number of unknown forgery types is known *a priori*. To address these challenges, we propose a Confidence-Aware Asymmetric Learning (CAL) framework, which adaptively balances model confidence across known and novel forgery types. CAL mainly consists of two components: Confidence-Aware Consistency Regularization (CCR) and Asymmetric Confidence Reinforcement (ACR). CCR mitigates pseudo-label bias by dynamically scaling sample losses based on normalized confidence, gradually shifting the training focus from high- to low-confidence samples. ACR complements this by separately calibrating confidence for known and novel classes through selective learning on high-confidence samples, guided by their confidence gap. Together, CCR and ACR form a mutually reinforcing loop that significantly improves the model's OW-DFA performance. Moreover, we introduce a Dynamic Prototype Pruning (DPP) strategy that automatically estimates the number of novel forgery types in a coarse-to-fine manner, removing the need for unrealistic prior assumptions and enhancing the scalability of our methods to real-world OW-DFA scenarios. Extensive experiments on the standard OW-DFA benchmark and a newly extended benchmark incorporating advanced manipulations demonstrate that CAL consistently outperforms previous methods, achieving new state-of-the-art performance on both known and novel forgery attribution.
翻译:合成人脸图像的激增加剧了对鲁棒开放世界深度伪造溯源(OW-DFA)的需求,该任务旨在利用已知类型的标注数据和包含已知与新型伪造类型的混合未标注数据,对已知和未知伪造进行溯源。然而,现有的OW-DFA方法面临两个关键局限:1)置信度偏斜导致对新型伪造生成不可靠的伪标签,从而引发训练偏差。2)一个不切实际的假设,即未知伪造类型的数量是*先验已知*的。为应对这些挑战,我们提出了一个置信度感知的非对称学习框架,该框架自适应地平衡模型在已知与新型伪造类型间的置信度。CAL主要由两个组件构成:置信度感知一致性正则化与不对称置信度增强。CCR通过基于归一化置信度动态缩放样本损失来缓解伪标签偏差,逐步将训练重点从高置信度样本转移到低置信度样本。ACR则通过对高置信度样本进行选择性学习,并以其置信度差距为指导,分别校准已知类与新型类的置信度,从而对CCR形成补充。CCR与ACR共同构成一个相互增强的循环,显著提升了模型的OW-DFA性能。此外,我们提出了一种动态原型剪枝策略,该策略以从粗到精的方式自动估计新型伪造类型的数量,从而消除了不切实际的先验假设需求,并增强了我们方法在现实世界OW-DFA场景中的可扩展性。在标准OW-DFA基准以及一个融合了先进篡改技术的新扩展基准上进行的大量实验表明,CAL始终优于先前的方法,在已知和新型伪造溯源任务上均实现了新的最先进性能。