We introduce a new dataset for the emotional artificial intelligence research: identity-free video dataset for Micro-Gesture Understanding and Emotion analysis (iMiGUE). Different from existing public datasets, iMiGUE focuses on nonverbal body gestures without using any identity information, while the predominant researches of emotion analysis concern sensitive biometric data, like face and speech. Most importantly, iMiGUE focuses on micro-gestures, i.e., unintentional behaviors driven by inner feelings, which are different from ordinary scope of gestures from other gesture datasets which are mostly intentionally performed for illustrative purposes. Furthermore, iMiGUE is designed to evaluate the ability of models to analyze the emotional states by integrating information of recognized micro-gesture, rather than just recognizing prototypes in the sequences separately (or isolatedly). This is because the real need for emotion AI is to understand the emotional states behind gestures in a holistic way. Moreover, to counter for the challenge of imbalanced sample distribution of this dataset, an unsupervised learning method is proposed to capture latent representations from the micro-gesture sequences themselves. We systematically investigate representative methods on this dataset, and comprehensive experimental results reveal several interesting insights from the iMiGUE, e.g., micro-gesture-based analysis can promote emotion understanding. We confirm that the new iMiGUE dataset could advance studies of micro-gesture and emotion AI.
翻译:我们引入了情感人工智能研究的新数据集:为微化理解和情感分析(iMIGUE)引入了没有身份的视频数据集。与现有的公共数据集不同,iMIGUE侧重于非语言的身体动作,而没有使用任何身份信息,而情感分析的主要研究则涉及敏感的生物鉴别数据,比如脸和言语。最重要的是,iMIGUE侧重于由情感驱动的微动,即由内心情感驱动的无意行为,这不同于一般的手势动作范围,而其他手势数据集大多是故意为说明目的进行的。此外,iMIGUE旨在评估模型分析情绪状态的能力,办法是整合公认的微进化图像信息,而不是仅仅单独(或孤立地)地识别序列中的原型。这是因为对情感分析的真正需要是以整体的方式理解手势背后的情感状态。此外,为了应对该数据集样本分布不平衡的挑战,我们提出了一种未超超超的学习方法,以便从若干基于微化图像的预视像序列中获取潜在表现的能力。我们系统地调查了这种数据分析结果。