Deep Learning and Machine Learning based models have become extremely popular in text processing and information retrieval. However, the non-linear structures present inside the networks make these models largely inscrutable. A significant body of research has focused on increasing the transparency of these models. This article provides a broad overview of research on the explainability and interpretability of natural language processing and information retrieval methods. More specifically, we survey approaches that have been applied to explain word embeddings, sequence modeling, attention modules, transformers, BERT, and document ranking. The concluding section suggests some possible directions for future research on this topic.
翻译:在文本处理和信息检索方面,基于深层学习和机器学习的模型已变得极为流行,然而,网络内存在的非线性结构使这些模型基本上难以确定,大量研究的重点是提高这些模型的透明度,这一条广泛概述了关于自然语言处理和信息检索方法的解释和可解释性的研究,更具体地说,我们调查了用于解释文字嵌入、顺序建模、注意模块、变压器、BERT和文件排序的方法。