A number of studies have demonstrated the efficacy of deep learning convolutional neural network (CNN) models for ocular-based user recognition in mobile devices. However, these high-performing networks have enormous space and computational complexity due to the millions of parameters and computations involved. These requirements make the deployment of deep learning models to resource-constrained mobile devices challenging. To this end, only a handful of studies based on knowledge distillation and patch-based models have been proposed to obtain compact size CNN models for ocular recognition in the mobile environment. In order to further advance the state-of-the-art, this study for the first time evaluates five neural network pruning methods and compares them with the knowledge distillation method for on-device CNN inference and mobile user verification using ocular images. Subject-independent analysis on VISOB and UPFR-Periocular datasets suggest the efficacy of layerwise magnitude-based pruning at a compression rate of 8 for mobile ocular-based authentication using ResNet50 as the base model. Further, comparison with the knowledge distillation suggests the efficacy of knowledge distillation over pruning methods in terms of verification accuracy and the real-time inference measured as deep feature extraction time on five mobile devices, namely, iPhone 6, iPhone X, iPhone XR, iPad Air 2 and iPad 7th Generation.
翻译:一些研究显示,在移动设备中,深学习进化神经网络(CNN)模型对基于视觉的用户进行视觉识别是有效的;然而,由于涉及数以百万计的参数和计算,这些高性能网络具有巨大的空间和计算复杂性。这些要求使得在资源紧张的移动设备中部署深学习模型具有挑战性。为此,根据知识蒸馏和基于补丁的模型,只提议进行少量研究,以便在移动环境中获得用于视觉识别的紧凑规模CNN模型;为了进一步推进最新技术,本项研究首次评估了5种神经网络运行方法,并将它们与使用视觉图像在CNN设备上进行推断和移动用户核查的知识蒸馏方法进行比较。根据对VISOB和UPFR-血管数据集进行的专题分析表明,在使用ResNet50作为基础模型的基于8级压缩速度进行移动望远镜认证时,基于7PP50的压压压率运行。此外,与知识蒸馏法显示,XPero 即测测测测测算的精确度,即测测测测测测深层的IPPPro 5号的精确度,即测测测测测测测测深层的IPPPro 方法。