How is knowledge of position-role mappings in natural language learned? We explore this question in a computational setting, testing whether a variety of well-performing pertained language models (BERT, RoBERTa, and DistilBERT) exhibit knowledge of these mappings, and whether this knowledge persists across alternations in syntactic, structural, and lexical alternations. In Experiment 1, we show that these neural models do indeed recognize distinctions between theme and recipient roles in ditransitive constructions, and that these distinct patterns are shared across construction type. We strengthen this finding in Experiment 2 by showing that fine-tuning these language models on novel theme- and recipient-like tokens in one paradigm allows the models to make correct predictions about their placement in other paradigms, suggesting that the knowledge of these mappings is shared rather than independently learned. We do, however, observe some limitations of this generalization when tasks involve constructions with novel ditransitive verbs, hinting at a degree of lexical specificity which underlies model performance.
翻译:如何在自然语言中学习定位功能绘图知识? 我们在计算环境中探索这一问题,测试各种表现良好的语言模型(BERT、ROBERTA和DistilBERT)是否展示了这些绘图知识,以及这种知识是否存在于合成、结构和词汇交替的交替状态中。在实验1中,我们证明这些神经模型确实认识到主题和接受方在二流构造中的角色之间的区别,这些不同的模式在建筑类型中是共享的。我们通过实验2中显示这些语言模型在新颖主题和接受方相似的符号上进行微调,使这些模型能够正确预测这些模型在其他模式中的位置,表明这些绘图的知识是共享的,而不是独立学习的。然而,当任务涉及建筑时,我们确实看到这种概括性的一些局限性,我们暗示了模型性能所依据的某种程度的通俗性特征。