The field of natural language understanding has experienced exponential progress in the last few years, with impressive results in several tasks. This success has motivated researchers to study the underlying knowledge encoded by these models. Despite this, attempts to understand their semantic capabilities have not been successful, often leading to non-conclusive, or contradictory conclusions among different works. Via a probing classifier, we extract the underlying knowledge graph of nine of the most influential language models of the last years, including word embeddings, text generators, and context encoders. This probe is based on concept relatedness, grounded on WordNet. Our results reveal that all the models encode this knowledge, but suffer from several inaccuracies. Furthermore, we show that the different architectures and training strategies lead to different model biases. We conduct a systematic evaluation to discover specific factors that explain why some concepts are challenging. We hope our insights will motivate the development of models that capture concepts more precisely.
翻译:自然语言理解领域在过去几年里取得了巨大进步,在一些任务中取得了令人印象深刻的成果。 这一成功激励研究人员研究这些模型所编码的基本知识。 尽管如此,试图理解其语义能力的努力一直没有成功,往往导致不同作品之间产生非结论性或相互矛盾的结论。 通过进行筛选的分类,我们提取了过去几年中九种最有影响力语言模型的基本知识图,包括单词嵌入、文本生成器和背景编码器。这一探测基于WordNet的概念关联性。 我们的结果显示,所有模型都为这种知识编码,但有一些不准确之处。 此外,我们显示,不同的结构和培训战略导致不同的模式偏差。我们进行系统评估,以发现解释某些概念为何具有挑战性的具体因素。我们希望我们的洞察力将推动更准确地捕捉到概念的模型的发展。