Fairness and environmental impact are important research directions for the sustainable development of artificial intelligence. However, while each topic is an active research area in natural language processing (NLP), there is a surprising lack of research on the interplay between the two fields. This lacuna is highly problematic, since there is increasing evidence that an exclusive focus on fairness can actually hinder environmental sustainability, and vice versa. In this work, we shed light on this crucial intersection in NLP by (1) investigating the efficiency of current fairness approaches through surveying example methods for reducing unfair stereotypical bias from the literature, and (2) evaluating a common technique to reduce energy consumption (and thus environmental impact) of English NLP models, knowledge distillation (KD), for its impact on fairness. In this case study, we evaluate the effect of important KD factors, including layer and dimensionality reduction, with respect to: (a) performance on the distillation task (natural language inference and semantic similarity prediction), and (b) multiple measures and dimensions of stereotypical bias (e.g., gender bias measured via the Word Embedding Association Test). Our results lead us to clarify current assumptions regarding the effect of KD on unfair bias: contrary to other findings, we show that KD can actually decrease model fairness.
翻译:公平性和环境影响是人造情报可持续发展的重要研究方向,然而,尽管每个专题都是自然语言处理的一个积极研究领域(NLP),但令人惊讶的是,对这两个领域之间的相互作用缺乏研究。这一空白问题很大,因为越来越多的证据表明,完全注重公平实际上会阻碍环境可持续性,反之亦然。在这项工作中,我们通过:(1) 通过调查减少文献中不公平的定型偏见的典型方法,调查当前公平做法的效率,以及(2) 评价减少英国NLP模型的能源消耗(并因此对环境的影响)的共同技术、知识蒸馏(KD),以了解其对公平的影响。在本案例研究中,我们评估了重要的KD因素的影响,包括层和维度的减少,涉及:(a) 蒸馏任务(自然语言推论和语义性相似性预测)的绩效,以及(b) 定型偏见的多重措施和层面(例如,通过Word Empeding Asociation),通过知识蒸馏(KD)的模型,我们实际上能够澄清目前关于KD的不公平的假设。