Deep learning-based face recognition models follow the common trend in deep neural networks by utilizing full-precision floating-point networks with high computational costs. Deploying such networks in use-cases constrained by computational requirements is often infeasible due to the large memory required by the full-precision model. Previous compact face recognition approaches proposed to design special compact architectures and train them from scratch using real training data, which may not be available in a real-world scenario due to privacy concerns. We present in this work the QuantFace solution based on low-bit precision format model quantization. QuantFace reduces the required computational cost of the existing face recognition models without the need for designing a particular architecture or accessing real training data. QuantFace introduces privacy-friendly synthetic face data to the quantization process to mitigate potential privacy concerns and issues related to the accessibility to real training data. Through extensive evaluation experiments on seven benchmarks and four network architectures, we demonstrate that QuantFace can successfully reduce the model size up to 5x while maintaining, to a large degree, the verification performance of the full-precision model without accessing real training datasets.
翻译:深层学习的面部识别模型与深神经网络的共同趋势相一致,利用全精度浮点网络,计算成本高。由于全精度模型需要大量的内存,将这种网络部署在受计算要求限制的用箱中往往不可行。 以往的契约面部识别方法建议使用真实的培训数据来设计特别的紧凑结构,并用真实的培训数据从零开始对其进行培训。我们在此工作中介绍了基于低比精确度格式模型量化的QuantFace解决方案。QuantFace降低了现有面部识别模型所需的计算费用,而无需设计特定架构或获取实际培训数据。QuantFace采用方便隐私的合成数据,以量化过程为方便,减少潜在的隐私关切和与实际培训数据获取有关的问题。通过对七个基准和四个网络结构进行广泛的评价实验,我们证明QuantFace能够成功地将模型的大小减少到5x,同时将不需大量地保持全面访问数据模型的核查性能。