Pre-trained language models learn large amounts of knowledge from their training corpus, while the memorized facts could become outdated over a few years. Model editing aims to make post-hoc updates on specific facts in a model while leaving irrelevant knowledge unchanged. However, existing work studies only the monolingual scenario. In this paper, we focus on cross-lingual model editing. Firstly, we propose the definition and metrics of the cross-lingual model editing, where updates in a single language should take effect in the others as well. Next, we propose a simple framework to convert a monolingual model editing approach to its cross-lingual variant using the parallel corpus. Experiments show that such an approach outperforms monolingual baselines by a large margin. Furthermore, we propose language anisotropic editing to improve cross-lingual editing by estimating parameter importance for each language. Experiments reveal that language anisotropic editing decreases the editing failing rate by another $26\%$ relatively.
翻译:培训前语言模式从培训中学习大量知识,而记忆化事实可能会在几年内过时。模式编辑的目的是在模型中对具体事实进行热后更新,而使不相关的知识保持不变。但是,现有工作研究只对单语种情况进行。在本文中,我们侧重于跨语言模式编辑。首先,我们提议跨语言模式编辑的定义和衡量标准,用单一语言进行更新,在其他语言中也应生效。接着,我们提议一个简单的框架,利用平行体将单一语言模式编辑方法转换为跨语言版本。实验表明,这种方法大大优于单语基线。此外,我们提议语言异语编辑,通过估计每种语言的参数重要性,改进跨语言编辑。实验显示,语言异口式编辑将编辑失败率相对减少26美元。