The logical negation property (LNP), which implies generating different predictions for semantically opposite inputs, is an important property that a trustworthy language model must satisfy. However, much recent evidence shows that large-size pre-trained language models (PLMs) do not satisfy this property. In this paper, we perform experiments using probing tasks to assess PLM's LNP understanding. Unlike previous studies that only examined negation expressions, we expand the boundary of the investigation to lexical semantics. Through experiments, we observe that PLMs violate the LNP frequently. To alleviate the issue, we propose a novel intermediate training task, names meaning-matching, designed to directly learn a meaning-text correspondence, instead of relying on the distributional hypothesis. Through multiple experiments, we find that the task enables PLMs to learn lexical semantic information. Also, through fine-tuning experiments on 7 GLUE tasks, we confirm that it is a safe intermediate task that guarantees a similar or better performance of downstream tasks. Finally, we observe that our proposed approach outperforms our previous counterparts despite its time and resource efficiency.
翻译:逻辑否定性财产(LNP)意味着对语义截然相反的投入产生不同的预测,这是值得信赖的语言模式必须满足的重要属性。然而,最近许多证据表明,大型的预先培训语言模型(PLM)不能满足这一属性。在本文中,我们用测试任务来评估PLM对LNP的理解。与以往只研究否定性表达方式的研究不同,我们把调查的界限扩大到词汇性语义。通过实验,我们发现PLM经常违反LNP。为了缓解这一问题,我们提议了一项新型的中级培训任务,即名称含义匹配,目的是直接学习文字文字信函,而不是依赖分布式假设。我们通过多次实验发现,任务使PLMS能够学习词汇性语义信息。此外,通过对7 GLUE任务进行微调实验,我们确认这是一项安全的中间任务,可以保证类似或更好地执行下游任务。最后,我们发现,我们提出的方法超越了我们先前的对应方,尽管有时间和资源效率。