A characteristic feature of human semantic memory is its ability to not only store and retrieve the properties of concepts observed through experience, but to also facilitate the inheritance of properties (can breathe) from superordinate concepts (animal) to their subordinates (dog) -- i.e. demonstrate property inheritance. In this paper, we present COMPS, a collection of minimal pair sentences that jointly tests pre-trained language models (PLMs) on their ability to attribute properties to concepts and their ability to demonstrate property inheritance behavior. Analyses of 22 different PLMs on COMPS reveal that they can easily distinguish between concepts on the basis of a property when they are trivially different, but find it relatively difficult when concepts are related on the basis of nuanced knowledge representations. Furthermore, we find that PLMs can demonstrate behavior consistent with property inheritance to a great extent, but fail in the presence of distracting information, which decreases the performance of many models, sometimes even below chance. This lack of robustness in demonstrating simple reasoning raises important questions about PLMs' capacity to make correct inferences even when they appear to possess the prerequisite knowledge.
翻译:人类语义记忆的一个特征是它不仅能够储存和检索通过经验观察到的概念的属性,而且能够促进财产的继承(能够呼吸)从超大概念(动物)到下属(狗) -- -- 即展示财产继承。在本文中,我们介绍了一套最小的配对句,共同测试了经过训练的语言模型(PLMS)将特性归因于概念的能力和显示财产继承行为的能力。对22个不同的COMPS中PLM的分析表明,当一个财产存在微小差异时,它们可以很容易地区分这些概念,但当概念根据细微知识的表述与概念相关时,发现这些概念相对困难。此外,我们认为,PLMMS可以在很大程度上展示与财产继承相一致的行为,但在出现分散注意力的信息时,这降低了许多模型的性能,有时甚至低于机会。 缺乏有力的论证简单推理的能力,使人对PLMs是否有能力作出正确的推论提出了重要问题,即使他们似乎具备先决条件的知识。