With the emergence of GAN, face forgery technologies have been heavily abused. Achieving accurate face forgery detection is imminent. Inspired by remote photoplethysmography (rPPG) that PPG signal corresponds to the periodic change of skin color caused by heartbeat in face videos, we observe that despite the inevitable loss of PPG signal during the forgery process, there is still a mixture of PPG signals in the forgery video with a unique rhythmic pattern depending on its generation method. Motivated by this key observation, we propose a framework for face forgery detection and categorization consisting of: 1) a Spatial-Temporal Filtering Network (STFNet) for PPG signals filtering, and 2) a Spatial-Temporal Interaction Network (STINet) for constraint and interaction of PPG signals. Moreover, with insight into the generation of forgery methods, we further propose intra-source and inter-source blending to boost the performance of the framework. Overall, extensive experiments have proved the superiority of our method.
翻译:随着GAN的出现,面部伪造技术被严重滥用。 准确的面部伪造检测即将实现。 在远程光谱扫描(rPPG)的启发下,PPG信号与脸部视频心跳的心跳的皮肤定期变化相对应,我们注意到,尽管在伪造过程中,PPPG信号不可避免地丢失了,但伪造视频中仍有一种PPPG信号的混合,其独特的有节奏模式取决于其生成方法。受这一关键观察的驱动,我们提出了一个面部伪造检测和分类框架,包括:(1) 用于PPPG信号过滤的空间时空过滤网络(STFNet),和(2) 用于限制PPG信号和互动的空间时空互动网络(STINet)。此外,随着对伪造方法的生成的深入了解,我们进一步提议了内部和源间混合,以提高框架的性能。总体而言,广泛的实验证明了我们的方法的优势。