Are the predictions of humans and language models affected by similar things? Research suggests that while comprehending language, humans make predictions about upcoming words, with more predictable words being processed more easily. However, evidence also shows that humans display a similar processing advantage for highly anomalous words when these words are semantically related to the preceding context or to the most probable continuation. Using stimuli from 3 psycholinguistic experiments, we find that this is also almost always also the case for 8 contemporary transformer language models (BERT, ALBERT, RoBERTa, XLM-R, GPT-2, GPT-Neo, GPT-J, and XGLM). We then discuss the implications of this phenomenon for our understanding of both human language comprehension and the predictions made by language models.
翻译:人类和语言模型的预测是否受到类似事物的影响?研究表明,在理解语言的同时,人类对即将到来的词汇作出预测,而更可预测的词汇则更容易处理。然而,证据还表明,当这些词汇与前一种或最可能延续的语义相关时,人类对于非常反常的词汇也表现出类似的处理优势。我们利用3种精神语言实验的刺激,发现这几乎也总是适用于8种当代变异语言模型(BERT、ALBERT、ROBERTA、XLM-R、GPT-2、GPT-Neo、GPT-J和XGLM)。 然后我们讨论了这一现象对我们理解人类语言理解和语言模型预测的影响。