The remarkable progress in Natural Language Processing (NLP) brought about by deep learning, particularly with the recent advent of large pre-trained neural language models, is brought into scrutiny as several studies began to discuss and report potential biases in NLP applications. Bias in NLP is found to originate from latent historical biases encoded by humans into textual data which gets perpetuated or even amplified by NLP algorithm. We present a survey to comprehend bias in large pre-trained language models, analyze the stages at which they occur in these models, and various ways in which these biases could be quantified and mitigated. Considering wide applicability of textual affective computing based downstream tasks in real-world systems such as business, healthcare, education, etc., we give a special emphasis on investigating bias in the context of affect (emotion) i.e., Affective Bias, in large pre-trained language models. We present a summary of various bias evaluation corpora that help to aid future research and discuss challenges in the research on bias in pre-trained language models. We believe that our attempt to draw a comprehensive view of bias in pre-trained language models, and especially the exploration of affective bias will be highly beneficial to researchers interested in this evolving field.
翻译:由于一些研究开始讨论和报告自然语言应用中潜在的偏差,因此,对通过深层次学习,特别是最近出现大量经过训练的神经语言模型而在自然语言处理方面取得的显著进展进行了仔细审查。发现自然语言应用中的Bias起源于人类潜伏的历史偏见,编成由人类编成的文本数据,由自然语言学算法延续或甚至放大。我们提出一份调查,以理解大型经过训练的语文模型中的偏见,分析这些模式中出现的偏见的各阶段,以及这些偏见可以量化和减轻的各种方式。考虑到基于文字影响计算在实际世界系统,例如商业、保健、教育等下游系统中的广泛适用性,我们特别强调调查影响(情感)背景下的偏见,即Affective Bias,以经过训练的大型语言模型为基础。我们提出一份各种偏见评价公司的摘要,以帮助今后的研究,并讨论这些模式中存在的偏见问题。我们认为,我们试图在经过训练前语言模型中的偏见方面提出全面的看法,将极大地影响正在演变的研究人员。