Face Anti-Spoofing (FAS) is essential to secure face recognition systems and has been extensively studied in recent years. Although deep neural networks (DNNs) for the FAS task have achieved promising results in intra-dataset experiments with similar distributions of training and testing data, the DNNs' generalization ability is limited under the cross-domain scenarios with different distributions of training and testing data. To improve the generalization ability, recent hybrid methods have been explored to extract task-aware handcrafted features (e.g., Local Binary Pattern) as discriminative information for the input of DNNs. However, the handcrafted feature extraction relies on experts' domain knowledge, and how to choose appropriate handcrafted features is underexplored. To this end, we propose a learnable network to extract Meta Pattern (MP) in our learning-to-learn framework. By replacing handcrafted features with the MP, the discriminative information from MP is capable of learning a more generalized model. Moreover, we devise a two-stream network to hierarchically fuse the input RGB image and the extracted MP by using our proposed Hierarchical Fusion Module (HFM). We conduct comprehensive experiments and show that our MP outperforms the compared handcrafted features. Also, our proposed method with HFM and the MP can achieve state-of-the-art performance on two different domain generalization evaluation benchmarks.
翻译:虽然用于FAS任务的深神经网络(DNN)在以类似的培训和测试数据分布的类似数据进行的数据内部实验中取得了有希望的成果,但DNN的概括化能力在交叉情景下有限,因为培训和测试数据分布不同。为了提高概括化能力,已经探索了最近的混合方法,以提取有任务手工制作的特征(例如,地方二进制模式)作为DNN投入的歧视性信息。然而,手工制作的特征提取依靠专家的域知识,如何选择适当的手工制作特征也没有得到充分探讨。为此,我们提议建立一个可学习的网络,在我们从学习到测试的框架中提取元模式。通过用MP取代手工制作的特征,MP的歧视性信息能够学习一个更普遍的模型。此外,我们设计了一个双流网络,将输入的 RGB 图像和提取的MP 依据专家域知识,如何选择适当的手工制作的特征也没有得到充分探讨。为此,我们提出了一个可学习到学习的模型,从我们总体的模型和MFMFMS 展示了我们所提议的总体模式。