Over the last few decades, artificial intelligence research has made tremendous strides, but it still heavily relies on fixed datasets in stationary environments. Continual learning is a growing field of research that examines how AI systems can learn sequentially from a continuous stream of linked data in the same way that biological systems do. Simultaneously, fake media such as deepfakes and synthetic face images have emerged as significant to current multimedia technologies. Recently, numerous method has been proposed which can detect deepfakes with high accuracy. However, they suffer significantly due to their reliance on fixed datasets in limited evaluation settings. Therefore, in this work, we apply continuous learning to neural networks' learning dynamics, emphasizing its potential to increase data efficiency significantly. We propose Continual Representation using Distillation (CoReD) method that employs the concept of Continual Learning (CoL), Representation Learning (ReL), and Knowledge Distillation (KD). We design CoReD to perform sequential domain adaptation tasks on new deepfake and GAN-generated synthetic face datasets, while effectively minimizing the catastrophic forgetting in a teacher-student model setting. Our extensive experimental results demonstrate that our method is efficient at domain adaptation to detect low-quality deepfakes videos and GAN-generated images from several datasets, outperforming the-state-of-art baseline methods.
翻译:在过去几十年里,人工智能研究取得了巨大的进步,但它仍然严重依赖固定环境中的固定数据集。不断学习是一个日益扩大的研究领域,它研究AI系统如何以生物系统同样的方式从连续的链接数据流中相继学习。与此同时,深假和合成面像等假媒体已经出现,对目前的多媒体技术具有重要意义。最近,提出了许多方法,能够以高度精确的方式探测深度假象。然而,由于在有限的评价环境中依赖固定数据集,它们遭受了巨大损失。因此,我们在这项工作中,对神经网络的学习动态进行不断学习,强调其大大提高数据效率的潜力。我们建议采用不断演示(CoReD)方法,采用持续学习(CoL)、演示学习(REL)和知识蒸馏(KD)的概念。我们设计了CORED,以便在新的深假和GAN生成的合成面数据集上执行连续的域适应任务,同时有效地将灾难性的遗忘在教师研究模型设定中,同时强调其显著提高数据效率的潜力。我们广泛的实验结果显示,从低质量的GAN模型到低质量的模型显示我们的有效域方法。