Logical approaches to representing language have developed and evaluated computational models of quantifier words since the 19th century, but today's NLU models still struggle to capture their semantics. We rely on Generalized Quantifier Theory for language-independent representations of the semantics of quantifier words, to quantify their contribution to the errors of NLU models. We find that quantifiers are pervasive in NLU benchmarks, and their occurrence at test time is associated with performance drops. Multilingual models also exhibit unsatisfying quantifier reasoning abilities, but not necessarily worse for non-English languages. To facilitate directly-targeted probing, we present an adversarial generalized quantifier NLI task (GQNLI) and show that pre-trained language models have a clear lack of robustness in generalized quantifier reasoning.
翻译:自19世纪以来,代表语言的逻辑方法已经开发并评估了量化词的计算模型,但今天的NLU模型仍然难以捕捉其语义。我们依靠通用量化词理论来独立语言表达量化词的语义,以量化其对NLU模型错误的贡献。我们发现,量化词在NLU基准中很普遍,测试时的出现与性能下降有关。多语言模型也表现出不满意量化词的推理能力,但对于非英语语言来说不一定更差。为了便利直接有针对性的标注,我们提出了一个对抗性通用量化词NLI任务(GQNLI),并表明预先培训的语言模型在通用量化推理中明显缺乏稳健性。