Face presentation attack detection (PAD) is an essential measure to protect face recognition systems from being spoofed by malicious users and has attracted great attention from both academia and industry. Although most of the existing methods can achieve desired performance to some extent, the generalization issue of face presentation attack detection under cross-domain settings (e.g., the setting of unseen attacks and varying illumination) remains to be solved. In this paper, we propose a novel framework based on asymmetric modality translation for face presentation attack detection in bi-modality scenarios. Under the framework, we establish connections between two modality images of genuine faces. Specifically, a novel modality fusion scheme is presented that the image of one modality is translated to the other one through an asymmetric modality translator, then fused with its corresponding paired image. The fusion result is fed as the input to a discriminator for inference. The training of the translator is supervised by an asymmetric modality translation loss. Besides, an illumination normalization module based on Pattern of Local Gravitational Force (PLGF) representation is used to reduce the impact of illumination variation. We conduct extensive experiments on three public datasets, which validate that our method is effective in detecting various types of attacks and achieves state-of-the-art performance under different evaluation protocols.
翻译:面部攻击检测(PAD)是保护面部识别系统不被恶意用户蒙骗的一项基本措施,吸引了学术界和工业界的极大关注。虽然大多数现有方法可以在一定程度上达到预期的性能,但在跨领域环境下(例如,设置隐形袭击和各种光化)脸部攻击检测的普遍化问题仍有待解决。在本文件中,我们提议了一个基于不对称模式翻译的新框架,用于在双调情景下对面部攻击检测。在这个框架内,我们建立了两种真实面部模式图像之间的联系。具体地说,一种模式的新型融合计划通过不对称模式翻译将一种模式的形象转化为另一种模式,然后与相应的配对图像相结合。融合的结果作为歧视因素提供给了推断者。翻译的培训受到不对称模式翻译损失的监督。此外,基于当地重力模式(PLGF)代表的照明正常化模块被用于减少错误变异的影响。我们对三种公共袭击模式进行了广泛的实验,根据不同的程序,在三种公共袭击模式下,我们测量了不同的性评估方法。我们根据不同的程序,对三种公共袭击进行了广泛的实验,根据不同的评估方法进行了不同的评估。