Face authentication usually utilizes deep learning models to verify users with high recognition accuracy. However, face authentication systems are vulnerable to various attacks that cheat the models by manipulating the digital counterparts of human faces. So far, lots of liveness detection schemes have been developed to prevent such attacks. Unfortunately, the attacker can still bypass these schemes by constructing wide-ranging sophisticated attacks. We study the security of existing face authentication services (e.g., Microsoft, Amazon, and Face++) and typical liveness detection approaches. Particularly, we develop a new type of attack, i.e., the low-cost 3D projection attack that projects manipulated face videos on a 3D face model, which can easily evade these face authentication services and liveness detection approaches. To this end, we propose FaceLip, a novel liveness detection scheme for face authentication, which utilizes unforgeable lip motion patterns built upon well-designed acoustic signals to enable a strong security guarantee. The unique lip motion patterns for each user are unforgeable because FaceLip verifies the patterns by capturing and analyzing the acoustic signals that are dynamically generated according to random challenges, which ensures that our signals for liveness detection cannot be manipulated. Specially, we develop robust algorithms for FaceLip to eliminate the impact of noisy signals in the environment and thus can accurately infer the lip motions at larger distances. We prototype FaceLip on off-the-shelf smartphones and conduct extensive experiments under different settings. Our evaluation with 44 participants validates the effectiveness and robustness of FaceLip.
翻译:面部认证通常使用深层次的学习模型来验证用户的认知准确度很高。 然而, 面部认证系统很容易受到各种攻击, 这些攻击通过操纵人脸的数字对等软件来欺骗模型。 到目前为止, 已经制定了许多活性检测计划来防止这些攻击。 不幸的是, 攻击者仍然可以通过建造广泛的复杂攻击来绕过这些计划。 我们研究现有面部认证服务( 如微软、亚马逊和面部++) 和典型的活性检测方法的安全性。 特别是, 我们开发了一种新的攻击类型, 即: 低成本的三维投影式袭击, 这些项目在3D面部模型上被操纵的视频很容易被欺骗。 这可以很容易地避开这些脸部认证服务和活性检测方法。 为此, 我们提议了FaceLip, 一个全新的活性检测系统, 利用精心设计的音响信号构建了难以想象的嘴部动作模式, 以便获得强有力的安全保障。 每个用户都难以想象的嘴部运动模式是无法想象的, 因为FaceLip通过捕捉到并分析动态生成的音象信号, 随机的挑战性挑战, 保证我们在更精确的行距定位上能测测测测到更精确的动作, 。