As social robots and other intelligent machines enter the home, artificial emotional intelligence (AEI) is taking center stage to address users' desire for deeper, more meaningful human-machine interaction. To accomplish such efficacious interaction, the next-generation AEI need comprehensive human emotion models for training. Unlike theory of emotion, which has been the historical focus in psychology, emotion models are a descriptive tools. In practice, the strongest models need robust coverage, which means defining the smallest core set of emotions from which all others can be derived. To achieve the desired coverage, we turn to word embeddings from natural language processing. Using unsupervised clustering techniques, our experiments show that with as few as 15 discrete emotion categories, we can provide maximum coverage across six major languages--Arabic, Chinese, English, French, Spanish, and Russian. In support of our findings, we also examine annotations from two large-scale emotion recognition datasets to assess the validity of existing emotion models compared to human perception at scale. Because robust, comprehensive emotion models are foundational for developing real-world affective computing applications, this work has broad implications in social robotics, human-machine interaction, mental healthcare, and computational psychology.
翻译:随着社交机器人和其他智能机器进入家庭,人工情感智能(AEI)正在进入中心舞台,以满足用户更深入、更有意义的人类机器互动的愿望。为了实现这种有效互动,下一代AEI需要全面的人类情感模型来进行培训。与情感理论不同,情感模型一直是心理学的历史焦点,是一种描述工具。在实践中,最强大的模型需要强有力的覆盖,这意味着定义最小的核心情感组群,所有其他人都可以从中产生。为了实现预期的覆盖,我们转向自然语言处理中的文字嵌入。我们用不受监督的组合技术,我们的实验显示,只有不到15个离散的情感类别,我们就能在六种主要语言(阿拉伯文、中文、英文、法文、西班牙文和俄文)中提供最大限度的覆盖。为了支持我们的调查结果,我们还检查了两个大规模情感识别数据集的注释,以评估现有情感模型相对于规模人类感知的正确性。因为坚固、全面的情感模型是开发真实世界影响计算应用程序的基础。这项工作在社会机器人、人体-机器互动、精神保健、计算和心理学方面有着广泛的影响。