Generating synthetic datasets for training face recognition models is challenging because dataset generation entails more than creating high fidelity images. It involves generating multiple images of same subjects under different factors (\textit{e.g.}, variations in pose, illumination, expression, aging and occlusion) which follows the real image conditional distribution. Previous works have studied the generation of synthetic datasets using GAN or 3D models. In this work, we approach the problem from the aspect of combining subject appearance (ID) and external factor (style) conditions. These two conditions provide a direct way to control the inter-class and intra-class variations. To this end, we propose a Dual Condition Face Generator (DCFace) based on a diffusion model. Our novel Patch-wise style extractor and Time-step dependent ID loss enables DCFace to consistently produce face images of the same subject under different styles with precise control. Face recognition models trained on synthetic images from the proposed DCFace provide higher verification accuracies compared to previous works by $6.11\%$ on average in $4$ out of $5$ test datasets, LFW, CFP-FP, CPLFW, AgeDB and CALFW. Code is available at https://github.com/mk-minchul/dcface
翻译:生成用于训练人脸识别模型的合成数据集具有挑战性,因为数据集生成不仅涉及创建高保真度图像,还涉及以符合真实图像条件分布的方式生成同一对象在不同因素下的多个图像(如姿势、光照、表情、衰老和遮挡的变化)。以前的研究已经研究了使用GAN或3D模型生成合成数据集的方法。在这项工作中,我们从组合对象外观(ID)和外部因素(样式)条件的角度来探讨这个问题。这两个条件提供了一种直接控制类内和类间变化的方式。为此,我们基于扩散模型提出了一个带有双条件人脸生成器(DCFace)。我们的新型分区样式提取器和时间步依赖ID损失使DCFace能够在准确控制下连续生成相同对象在不同样式下的人脸图像。使用来自建议的DCFace的合成图像训练的人脸识别模型在LFW、CFP-FP、CPLFW、AgeDB和CALFW等5个测试数据集中平均提高了6.11%的验证准确性。代码可在https://github.com/mk-minchul/dcface上获取。