The need for emotional inference from text continues to diversify as more and more disciplines integrate emotions into their theories and applications. These needs include inferring different emotion types, handling multiple languages, and different annotation formats. A shared model between different configurations would enable the sharing of knowledge and a decrease in training costs, and would simplify the process of deploying emotion recognition models in novel environments. In this work, we study how we can build a single model that can transition between these different configurations by leveraging multilingual models and Demux, a transformer-based model whose input includes the emotions of interest, enabling us to dynamically change the emotions predicted by the model. Demux also produces emotion embeddings, and performing operations on them allows us to transition to clusters of emotions by pooling the embeddings of each cluster. We show that Demux can simultaneously transfer knowledge in a zero-shot manner to a new language, to a novel annotation format and to unseen emotions. Code is available at https://github.com/gchochla/Demux-MEmo .
翻译:由于越来越多的学科将情感纳入其理论和应用中,对文本情感推断的需要继续多样化。这些需要包括推断不同的情感类型、处理多种语言和不同的批注格式。不同组合之间的共同模式将有利于共享知识和降低培训成本,并将简化在新环境下部署情感识别模型的过程。在这项工作中,我们研究如何建立一个单一模式,通过利用多种语言模型和Demux(一个基于变压器的模型,其投入包括感官的情感,使我们能够动态地改变模型预测的情绪),Demux还产生情感嵌入,并运行这些模式,使我们能够通过集中每个组群的嵌入,向情感集群过渡。我们表明Demux可以以零发方式同时将知识转移到新语言、新发的注意格式和看不见的情感。代码可在https://github.com/gchola/Demuux-MEmo中查阅。