In low-resource computing contexts, such as smartphones and other tiny devices, Both deep learning and machine learning are being used in a lot of identification systems. as authentication techniques. The transparent, contactless, and non-invasive nature of these face recognition technologies driven by AI has led to their meteoric rise in popularity in recent years. While they are mostly successful, there are still methods to get inside without permission by utilising things like pictures, masks, glasses, etc. In this research, we present an alternate authentication process that makes use of both facial recognition and the individual's distinctive temporal facial feature motions while they speak a password. Because the suggested methodology allows for a password to be specified in any language, it is not limited by language. The suggested model attained an accuracy of 96.1% when tested on the industry-standard MIRACL-VC1 dataset, demonstrating its efficacy as a reliable and powerful solution. In addition to being data-efficient, the suggested technique shows promising outcomes with as little as 10 positive video examples for training the model. The effectiveness of the network's training is further proved via comparisons with other combined facial recognition and lip reading models.
翻译:在低资源计算环境中,例如智能手机和其他小型设备,深度学习和机器学习被广泛应用于识别系统中,作为身份验证技术。这些人脸识别技术驱动了AI的透明、非接触和非侵入性特点在最近几年内迅速崛起,在大多数情况下都非常成功。虽然它们基本上是成功的,但仍然有方法可以未经许可地进入,例如利用像图片、口罩、眼镜等物品。在这项研究中,我们提出了一种另类的身份验证过程,它利用了面部识别和个体在说密码时产生的独特的时空面部特征动作。由于所建议的方法可以在任何语言中指定密码,因此它不受语言的限制。在行业标准的MIRACL-VC1数据集上进行测试时,所提出的模型达到了96.1%的准确率,证明它是一种可靠且强大的解决方案。除了数据效率高之外,所提出的技术在训练模型时只需要10个正向视频示例就可以显示出令人鼓舞的结果。通过与其他结合面部识别和唇语阅读模型进行比较,进一步证明了网络的训练效果。