Pretrained language models (PLM) achieve surprising performance on the Choice of Plausible Alternatives (COPA) task. However, whether PLMs have truly acquired the ability of causal reasoning remains a question. In this paper, we investigate the problem of semantic similarity bias and reveal the vulnerability of current COPA models by certain attacks. Previous solutions that tackle the superficial cues of unbalanced token distribution still encounter the same problem of semantic bias, even more seriously due to the utilization of more training data. We mitigate this problem by simply adding a regularization loss and experimental results show that this solution not only improves the model's generalization ability, but also assists the models to perform more robustly on a challenging dataset, BCOPA-CE, which has unbiased token distribution and is more difficult for models to distinguish cause and effect.
翻译:受过训练的语言模式(PLM)在选择可选替代物(COPA)的任务上取得了惊人的成绩。然而,PLMs是否真正具备了因果推理能力仍然是一个问题。在本文件中,我们调查了语义相似性偏见的问题,并揭示了目前COPA模式因某些攻击而容易受到伤害的问题。以前解决象征分配不平衡的肤浅线索的解决方案仍然遇到同样的语义偏见问题,由于使用更多的培训数据,这一问题更为严重。我们简单地通过增加正规化损失和实验结果来缓解这一问题。我们简单地通过增加一种正规化损失和实验结果来表明,这一解决方案不仅提高了模型的普及能力,而且还帮助模型在具有挑战性的数据集BCOPA-CE上更有力地发挥作用,BCOPA-CE是没有偏见的象征性分布,而且模型更难以辨别因果关系。