The creation of altered and manipulated faces has become more common due to the improvement of DeepFake generation methods. Simultaneously, we have seen detection models' development for differentiating between a manipulated and original face from image or video content. We have observed that most publicly available DeepFake detection datasets have limited variations, where a single face is used in many videos, resulting in an oversampled training dataset. Due to this, deep neural networks tend to overfit to the facial features instead of learning to detect manipulation features of DeepFake content. As a result, most detection architectures perform poorly when tested on unseen data. In this paper, we provide a quantitative analysis to investigate this problem and present a solution to prevent model overfitting due to the high volume of samples generated from a small number of actors. We introduce Face-Cutout, a data augmentation method for training Convolutional Neural Networks (CNN), to improve DeepFake detection. In this method, training images with various occlusions are dynamically generated using face landmark information irrespective of orientation. Unlike other general-purpose augmentation methods, it focuses on the facial information that is crucial for DeepFake detection. Our method achieves a reduction in LogLoss of 15.2% to 35.3% on different datasets, compared to other occlusion-based augmentation techniques. We show that Face-Cutout can be easily integrated with any CNN-based recognition model and improve detection performance.
翻译:由于深层假面板的改进,被改变和操纵的面孔的创建变得更为常见。 同时,我们看到了用于区分被操纵和原始面孔的探测模型与图像或视频内容的图像或视频内容的原始面孔的开发。我们观察到,大多数公开提供的深层假脸检测数据集的变异性有限,许多视频使用单一面孔,导致培训数据集过多。因此,深层神经网络往往过度适应面部特征,而不是学习探测深层假脸内容的操纵特征。因此,大多数检测结构在对隐蔽数据进行测试时表现不佳。在本文件中,我们提供了定量分析,以调查这一问题,并提出解决办法,防止由于从少数行为体生成的样本数量众多,模型出现过度匹配。我们引入了“面板”这一数据增强方法,用于培训进化神经网络(CNN),从而改进深层Fake的检测。在这种方法中,使用基于各种模型的图像可以动态生成,而无需定向即可轻易地改进面部标志性格信息。与其他通用增强方法不同,我们侧重于对深层F3的检测至关重要的面信息,我们采用不同方法,我们用“LO2”的升级方法,我们用不同的方式显示了“LALA”