Significant advances in deep learning have obtained hallmark accuracy rates for various computer vision applications. However, advances in deep generative models have also led to the generation of very realistic fake content, also known as deepfakes, causing a threat to privacy, democracy, and national security. Most of the current deepfake detection methods are deemed as a binary classification problem in distinguishing authentic images or videos from fake ones using two-class convolutional neural networks (CNNs). These methods are based on detecting visual artifacts, temporal or color inconsistencies produced by deep generative models. However, these methods require a large amount of real and fake data for model training and their performance drops significantly in cross dataset evaluation with samples generated using advanced deepfake generation techniques. In this paper, we thoroughly evaluate the efficacy of deep face recognition in identifying deepfakes, using different loss functions and deepfake generation techniques. Experimental investigations on challenging Celeb-DF and FaceForensics++ deepfake datasets suggest the efficacy of deep face recognition in identifying deepfakes over two-class CNNs and the ocular modality. Reported results suggest a maximum Area Under Curve (AUC) of 0.98 and an Equal Error Rate (EER) of 7.1% in detecting deepfakes using face recognition on the Celeb-DF dataset. This EER is lower by 16.6% compared to the EER obtained for the two-class CNN and the ocular modality on the Celeb-DF dataset. Further on the FaceForensics++ dataset, an AUC of 0.99 and EER of 2.04% were obtained. The use of biometric facial recognition technology has the advantage of bypassing the need for a large amount of fake data for model training and obtaining better generalizability to evolving deepfake creation techniques.
翻译:深层学习的重大进步已经为各种计算机视觉应用取得了标志性精确率;然而,深基因模型的进步也导致生成了非常现实的假内容,也称为深假,对隐私、民主和国家安全构成威胁。目前大部分深假探测方法被视为一个二进制分类问题,在使用两等的卷心神经网络(CNNs)将真实图像或视频与假图像和假图像区分开来时,使用双级的卷心神经网络(CNNs),这些方法基于探测深层基因模型产生的视觉艺术品、时间或颜色不一致。然而,这些方法还需要大量真实和假数据,供模型培训使用,用于模型培训的模型培训,但模型培训需要大量真实和假数据,用于模型培训,在交叉数据集评估中显著的交叉数据集中显著下降。