Progress in pre-trained language models has led to a surge of impressive results on downstream tasks for natural language understanding. Recent work on probing pre-trained language models uncovered a wide range of linguistic properties encoded in their contextualized representations. However, it is unclear whether they encode semantic knowledge that is crucial to symbolic inference methods. We propose a methodology for probing linguistic information for logical inference in pre-trained language model representations. Our probing datasets cover a list of linguistic phenomena required by major symbolic inference systems. We find that (i) pre-trained language models do encode several types of linguistic information for inference, but there are also some types of information that are weakly encoded, (ii) language models can effectively learn missing linguistic information through fine-tuning. Overall, our findings provide insights into which aspects of linguistic information for logical inference do language models and their pre-training procedures capture. Moreover, we have demonstrated language models' potential as semantic and background knowledge bases for supporting symbolic inference methods.
翻译:培训前语言模型的进展导致在自然语言理解的下游任务方面出现了令人印象深刻的成果。最近关于调查培训前语言模型的工作发现,在背景描述中,语言特性很广,但尚不清楚这些模型是否编码了对象征性推理方法至关重要的语义知识。我们提出了一种方法,用于在培训前语言模型表述中进行逻辑推论的语言信息。我们的探测数据集包括了主要象征性推断系统所要求的一系列语言现象。我们发现:(一) 培训前语言模型确实对几种类型的语言信息进行了编码,用于推断,但也发现有些类型的信息编码薄弱,(二) 语言模型可以通过微调有效地学习缺失的语言信息。总体而言,我们的调查结果为逻辑推理语言模型及其培训前程序采集的哪些语言信息提供了深刻的见解。此外,我们展示了语言模型作为支持象征性推理方法的语义和背景知识基础的潜力。