Deepfakes utilise Artificial Intelligence (AI) techniques to create synthetic media where the likeness of one person is replaced with another. There are growing concerns that deepfakes can be maliciously used to create misleading and harmful digital contents. As deepfakes become more common, there is a dire need for deepfake detection technology to help spot deepfake media. Present deepfake detection models are able to achieve outstanding accuracy (>90%). However, most of them are limited to within-dataset scenario, where the same dataset is used for training and testing. Most models do not generalise well enough in cross-dataset scenario, where models are tested on unseen datasets from another source. Furthermore, state-of-the-art deepfake detection models rely on neural network-based classification models that are known to be vulnerable to adversarial attacks. Motivated by the need for a robust deepfake detection model, this study adapts metamorphic testing (MT) principles to help identify potential factors that could influence the robustness of the examined model, while overcoming the test oracle problem in this domain. Metamorphic testing is specifically chosen as the testing technique as it fits our demand to address learning-based system testing with probabilistic outcomes from largely black-box components, based on potentially large input domains. We performed our evaluations on MesoInception-4 and TwoStreamNet models, which are the state-of-the-art deepfake detection models. This study identified makeup application as an adversarial attack that could fool deepfake detectors. Our experimental results demonstrate that both the MesoInception-4 and TwoStreamNet models degrade in their performance by up to 30\% when the input data is perturbed with makeup.
翻译:深假智能(AI) 使用深假检测技术来创建合成媒体, 将一个人的相似性替换为另一个人。 人们越来越担心深假可能会被恶意地用于制造误导和有害的数字内容。 随着深假发现技术变得更加常见, 极需要深假检测技术来帮助发现深假智能媒体。 当前的深假检测模型能够达到惊人的准确性( > 90 % ) 。 但是, 大部分模型局限于内部数据设置, 即同一数据集被用于培训和测试。 大多数模型在交叉数据设置假设中不够全面, 而在另一个来源的隐蔽数据集中测试模型。 此外, 最先进的深假假智能检测模型依赖于以神经网络为基础的分类模型来帮助发现深假智能的准确性能( > 90 % ) 。 多数模型在使用相同的数据设置时, 能够帮助识别影响被检测模型的稳健健性的潜在因素, 而在跨数据库中, 超越了测试的测试或缩略地, 将我们的数据检测结果作为大规模测试的结果, 我们的大规模测试是用黑箱进行。