Pretrained language models have demonstrated outstanding performance in many NLP tasks recently. However, their social intelligence, which requires commonsense reasoning about the current situation and mental states of others, is still developing. Towards improving language models' social intelligence, we focus on the Social IQA dataset, a task requiring social and emotional commonsense reasoning. Building on top of the pretrained RoBERTa and GPT2 models, we propose several architecture variations and extensions, as well as leveraging external commonsense corpora, to optimize the model for Social IQA. Our proposed system achieves competitive results as those top-ranking models on the leaderboard. This work demonstrates the strengths of pretrained language models, and provides viable ways to improve their performance for a particular task.
翻译:最近,经过培训的语言模式在许多国家语言方案任务中表现出了杰出的成绩,然而,他们的社会智慧要求对目前的情况和他人的精神状态进行常识推理,这种社会智慧仍在发展之中。为了改进语言模式的社会智能,我们把重点放在社会智囊团数据集上,这是一项需要社会和情感常识推理的任务。 在经过培训的罗贝塔和GPT2模型之上,我们提出了几项结构变异和扩展建议,并利用外部常识公司来优化社会智囊团的模式。 我们提议的系统取得了与领导板上那些顶级模式一样的竞争性成果。 这项工作展示了预先培训的语言模式的优势,并为改进特定任务的业绩提供了可行的方法。