Recursive noun phrases (NPs) have interesting semantic properties. For example, "my favorite new movie" is not necessarily my favorite movie, whereas "my new favorite movie" is. This is common sense to humans, yet it is unknown whether language models have such knowledge. We introduce the Recursive Noun Phrase Challenge (RNPC), a dataset of three textual inference tasks involving textual entailment and event plausibility comparison, precisely targeting the understanding of recursive NPs. When evaluated on RNPC, state-of-the-art Transformer models only perform around chance. Still, we show that such knowledge is learnable with appropriate data. We further probe the models for relevant linguistic features that can be learned from our tasks, including modifier semantic category and modifier scope. Finally, models trained on RNPC achieve strong zero-shot performance on an extrinsic Harm Detection evaluation task, showing the usefulness of the understanding of recursive NPs in downstream applications.
翻译:“我最喜欢的新电影”不一定是我最喜欢的电影,而“我最喜欢的电影”则是。这是人类的常识,但对于语言模型是否具有这种知识却还不清楚。我们引入了包含三种文字推论任务的数据集“recursive Noun Phrase Challenge”(RNPC),这三种文字推论任务涉及文字内含和事件可信任性比较,确切针对对循环性NP的理解。在对RNPC(最新变换器模型)的评估中,只有机会才能发挥作用。我们还是表明,这种知识可以通过适当的数据来学习。我们进一步探索可以从我们的任务中学习的相关语言特征模型,包括修饰语分类和修饰范围。最后,在RNPC(RNPC)培训的模型在外源损害探测评估任务上取得了强烈的零光性表现,显示了对下游应用中再生NP的理解的有用性。