Previous works show that Pre-trained Language Models (PLMs) can capture factual knowledge. However, some analyses reveal that PLMs fail to perform it robustly, e.g., being sensitive to the changes of prompts when extracting factual knowledge. To mitigate this issue, we propose to let PLMs learn the deterministic relationship between the remaining context and the masked content. The deterministic relationship ensures that the masked factual content can be deterministically inferable based on the existing clues in the context. That would provide more stable patterns for PLMs to capture factual knowledge than randomly masking. Two pre-training tasks are further introduced to motivate PLMs to rely on the deterministic relationship when filling masks. Specifically, we use an external Knowledge Base (KB) to identify deterministic relationships and continuously pre-train PLMs with the proposed methods. The factual knowledge probing experiments indicate that the continuously pre-trained PLMs achieve better robustness in factual knowledge capturing. Further experiments on question-answering datasets show that trying to learn a deterministic relationship with the proposed methods can also help other knowledge-intensive tasks.
翻译:先前的著作表明,受过训练的语言模型(PLM)可以捕捉事实知识,然而,一些分析显示,PLM未能强有力地发挥这种作用,例如,在获取事实知识时对提示的变化十分敏感。为了缓解这一问题,我们提议让PLM学会剩余背景和蒙面内容之间的决定性关系。确定性关系确保隐藏的事实内容能够根据背景中的现有线索加以确定。这将为PLM提供比随机遮蔽更稳定的模式,使PLM能够捕捉事实知识。在培训前进一步引入了两项任务,以激励PLM在填充面具时依赖确定性关系。具体地说,我们使用外部知识库(KB)来确定确定性关系,并持续对拟议方法进行预先培训的PLMs。事实实验表明,持续受过训练的PLMs在获取事实知识方面能够更加稳健。关于问答数据集的进一步实验表明,试图学习与拟议方法的确定性关系,也可以帮助其他知识密集型任务。