Due to the expensive data collection process, micro-expression datasets are generally much smaller in scale than those in other computer vision fields, rendering large-scale training less stable and feasible. In this paper, we aim to develop a protocol to automatically synthesize micro-expression training data that 1) are on a large scale and 2) allow us to train recognition models with strong accuracy on real-world test sets. Specifically, we discover three types of Action Units (AUs) that can well constitute trainable micro-expressions. These AUs come from real-world micro-expressions, early frames of macro-expressions, and the relationship between AUs and expression labels defined by human knowledge. With these AUs, our protocol then employs large numbers of face images with various identities and an existing face generation method for micro-expression synthesis. Micro-expression recognition models are trained on the generated micro-expression datasets and evaluated on real-world test sets, where very competitive and stable performance is obtained. The experimental results not only validate the effectiveness of these AUs and our dataset synthesis protocol but also reveal some critical properties of micro-expressions: they generalize across faces, are close to early-stage macro-expressions, and can be manually defined.
翻译:由于数据收集过程费用昂贵,微表层数据集的规模一般大大小于其他计算机视觉领域,使大规模培训不那么稳定和可行。在本文件中,我们的目标是制定一项协议,自动合成微表层培训数据,其中(1) 规模大,(2) 使我们能够在真实世界测试组中以高度精确的方式培训识别模型。具体地说,我们发现三种类型的行动单位(AUS)完全可以构成可培训的微观表现。这些AU来自真实世界的微观表现、宏观表达的早期框架,以及AUs与人类知识定义的表达标志之间的关系。有了这些AU,我们的协议然后使用大量具有不同身份的面部图像和现有的微表层合成面部生成方法。微表层识别模型在生成的微表层数据集方面得到了培训,并在真实世界测试组中进行了评估,这些测试非常具有竞争力和稳定性的性能。实验结果不仅验证了这些AUs和我们数据集综合协议的有效性,而且还揭示了某些关键特征:它们能够将手动的宏观面进行到早期分析。