Detecting forgery videos is highly desirable due to the abuse of deepfake. Existing detection approaches contribute to exploring the specific artifacts in deepfake videos and fit well on certain data. However, the growing technique on these artifacts keeps challenging the robustness of traditional deepfake detectors. As a result, the development of generalizability of these approaches has reached a blockage. To address this issue, given the empirical results that the identities behind voices and faces are often mismatched in deepfake videos, and the voices and faces have homogeneity to some extent, in this paper, we propose to perform the deepfake detection from an unexplored voice-face matching view. To this end, a voice-face matching method is devised to measure the matching degree of these two. Nevertheless, training on specific deepfake datasets makes the model overfit certain traits of deepfake algorithms. We instead, advocate a method that quickly adapts to untapped forgery, with a pre-training then fine-tuning paradigm. Specifically, we first pre-train the model on a generic audio-visual dataset, followed by the fine-tuning on downstream deepfake data. We conduct extensive experiments over three widely exploited deepfake datasets - DFDC, FakeAVCeleb, and DeepfakeTIMIT. Our method obtains significant performance gains as compared to other state-of-the-art competitors. It is also worth noting that our method already achieves competitive results when fine-tuned on limited deepfake data.
翻译:检测伪造视频是非常可取的,因为滥用了深假视频。 现有的检测方法有助于探索深假视频中的具体文物,并符合某些数据。 但是,这些文物上日益增长的技术对传统的深假探测器的强健性提出了挑战。 结果,这些方法的普及性发展达到了一个障碍。 为了解决这一问题, 经验结果显示, 声音和面孔背后的身份往往在深假视频中不匹配, 声音和面孔在某种程度上具有同质性。 在本文中, 我们提议从未探索的语音相匹配视图中进行深度的检测。 为了达到这一目的, 设计了一种声音相匹配方法来测量传统深假探测器的强度。 然而, 对这些方法的普及性化性能的培训使得这些方法过于适合深假算法的某些特征。 相反, 我们提倡一种方法, 快速适应未开发的伪造, 先是培训前的,然后是微调的范式。 具体地,我们首先将模型放在一个价值的普通视听数据集上,然后是精细的深深深深底数据测试。 我们的深深底的利用方法进行了广泛的数据。