In recent years, image and video manipulations with Deepfake have become a severe concern for security and society. Many detection models and datasets have been proposed to detect Deepfake data reliably. However, there is an increased concern that these models and training databases might be biased and, thus, cause Deepfake detectors to fail. In this work, we investigate the bias issue caused by public Deepfake datasets by (a) providing large-scale demographic and non-demographic attribute annotations of 47 different attributes for five popular Deepfake datasets and (b) comprehensively analysing AI-bias of three state-of-the-art Deepfake detection backbone models on these datasets. The investigation analyses the influence of a large variety of distinctive attributes (from over 65M labels) on the detection performance, including demographic (age, gender, ethnicity) and non-demographic (hair, skin, accessories, etc.) information. The results indicate that investigated databases lack diversity and, more importantly, show that the utilised Deepfake detection backbone models are strongly biased towards many investigated attributes. The Deepfake detection backbone methods, which are trained with biased datasets, might output incorrect detection results, thereby leading to generalisability, fairness, and security issues. We hope that the findings of this study and the annotation databases will help to evaluate and mitigate bias in future Deepfake detection techniques. The annotation datasets are publicly available.
翻译:近年来,深度伪造技术在图像和视频操纵中引起了严重的安全和社会关注。许多检测模型和数据集已经被提出,以可靠地检测深度伪造数据。然而,人们越来越担心,这些模型和训练数据库可能存在偏见,从而导致深度伪造检测器失灵。本文通过提供47种不同属性的大规模民族和非民族属性注释以及全面分析三种最先进的深度伪造检测主干模型在这些数据集上的AI偏见问题,研究了公共深度伪造数据集引起的偏见问题。研究分析了包括民族(年龄、性别、种族)和非民族(头发、皮肤、配饰等)信息在内的大量特征(超过6500万标签)对检测性能的影响。结果表明,调查的数据库缺乏多样性,更重要的是,所使用的深度伪造检测主干模型对许多调查属性存在强烈的偏见。使用偏见数据集训练的深度伪造检测主干方法可能会输出不正确的检测结果,从而导致通用性、公正性和安全性问题。我们希望本研究的发现和注释数据集能够帮助评估和缓解未来深度伪造检测技术中的偏见问题。这些注释数据集是公开可用的。