Motivations for methods in explainable artificial intelligence (XAI) often include detecting, quantifying and mitigating bias, and contributing to making machine learning models fairer. However, exactly how an XAI method can help in combating biases is often left unspecified. In this paper, we briefly review trends in explainability and fairness in NLP research, identify the current practices in which explainability methods are applied to detect and mitigate bias, and investigate the barriers preventing XAI methods from being used more widely in tackling fairness issues.
翻译:可解释的人工智能(XAI)方法的动态往往包括发现、量化和减少偏差,以及帮助使机器学习模式更加公平,然而,确切地说,XAI方法如何有助于打击偏差往往没有说明。 在本文中,我们简要回顾了NLP研究的解释性和公平性趋势,确定了目前应用解释性方法来发现和减少偏差的做法,并调查妨碍XAI方法被更广泛地用于解决公平问题的障碍。