Transformer-based language models have shown strong performance on an array of natural language understanding tasks. However, the question of how these models react to implicit meaning has been largely unexplored. We investigate this using the complement coercion phenomenon, which involves sentences like "The student finished the book about sailing" where the action "reading" is implicit. We compare LMs' surprisal estimates at various critical sentence regions in sentences with and without implicit meaning. Effects associated with recovering implicit meaning were found at a critical region other than where sentences minimally differ. We then use follow-up experiments to factor out potential confounds, revealing different perspectives that offer a richer and more accurate picture.
翻译:以变换器为基础的语言模型在一系列自然语言理解任务中表现出很强的表现。 但是,这些模型如何对隐含的含义作出反应的问题基本上没有得到探讨。 我们用补充胁迫现象来调查这一问题, 其中包括“ 学生完成了航海的书” 等句子, 即“ 阅读” 行动是隐含的。 我们比较了各个关键句子区域具有和不具有隐含含义的超常估计值。 与恢复隐含含义有关的影响是在一个关键区域发现的, 而不是在判决差别最小的区域。 我们接着使用后续实验来考虑潜在的混淆, 揭示出能提供更丰富和更准确的图片的不同观点 。