Does the effectiveness of neural language models derive entirely from accurate modeling of surface word co-occurrence statistics, or do these models represent and reason about the world they describe? In BART and T5 transformer language models, we identify contextual word representations that function as models of entities and situations as they evolve throughout a discourse. These neural representations have functional similarities to linguistic models of dynamic semantics: they support a linear readout of each entity's current properties and relations, and can be manipulated with predictable effects on language generation. Our results indicate that prediction in pretrained neural language models is supported, at least in part, by dynamic representations of meaning and implicit simulation of entity state, and that this behavior can be learned with only text as training data. Code and data are available at https://github.com/belindal/state-probes .
翻译:神经语言模型的有效性完全来自地表单词共发生统计的精确模型,还是这些模型代表了它们所描述的世界和理性?在BART和T5变压器语言模型中,我们确定了作为实体和各种情况在整个讨论中演进的模型的背景字表。这些神经语言模型在功能上与动态语义学模型有相似之处:它们支持对每个实体当前属性和关系的线性读取,并且可以对语言生成产生可预测的影响。我们的结果表明,在经过训练的神经语言模型中的预测至少部分得到实体状态的动态含义和隐性模拟的支持,而且只有作为培训数据的文本才能学习这种行为。代码和数据可在https://github.com/belindal/state-probes上查阅。