Deep-learning-based technologies such as deepfakes ones have been attracting widespread attention in both society and academia, particularly ones used to synthesize forged face images. These automatic and professional-skill-free face manipulation technologies can be used to replace the face in an original image or video with any target object while maintaining the expression and demeanor. Since human faces are closely related to identity characteristics, maliciously disseminated identity manipulated videos could trigger a crisis of public trust in the media and could even have serious political, social, and legal implications. To effectively detect manipulated videos, we focus on the position offset in the face blending process, resulting from the forced affine transformation of the normalized forged face. We introduce a method for detecting manipulated videos that is based on the trajectory of the facial region displacement. Specifically, we develop a virtual-anchor-based method for extracting the facial trajectory, which can robustly represent displacement information. This information was used to construct a network for exposing multidimensional artifacts in the trajectory sequences of manipulated videos that is based on dual-stream spatial-temporal graph attention and a gated recurrent unit backbone. Testing of our method on various manipulation datasets demonstrated that its accuracy and generalization ability is competitive with that of the leading detection methods.
翻译:深假人等深造技术在社会和学术界引起广泛关注,特别是用来合成假面像的深造技术。这些自动和专业的无技能面部操纵技术可以用来用任何目标对象取代原始图像或视频中的面部,同时保持表达和贬低。由于人的面孔与身份特征特征密切相关,恶意传播的身份操纵视频可能会引发公众信任媒体的危机,甚至可能产生严重的政治、社会和法律影响。为了有效检测被操纵视频,我们注重面部混合过程的方位抵消,这是由正常面部的被迫折叠改造造成的。我们采用一种方法,根据面部迁移的轨迹来探测被操纵的视频。具体地说,我们开发了一种基于虚拟安抚法的提取面部轨迹的方法,这可以有力地代表流离失所信息。这种信息被用来建立一个网络,在被操纵的视频的轨迹序列中暴露多层面的工艺品,其轨迹以双流空间-时空图形关注为基础,并有一个封闭的经常骨架为基础。我们以各种操纵能力测试方法的测试其是否具有竞争力。