Memorization presents a challenge for several constrained Natural Language Generation (NLG) tasks such as Neural Machine Translation (NMT), wherein the proclivity of neural models to memorize noisy and atypical samples reacts adversely with the noisy (web crawled) datasets. However, previous studies of memorization in constrained NLG tasks have only focused on counterfactual memorization, linking it to the problem of hallucinations. In this work, we propose a new, inexpensive algorithm for extractive memorization (exact training data generation under insufficient context) in constrained sequence generation tasks and use it to study extractive memorization and its effects in NMT. We demonstrate that extractive memorization poses a serious threat to NMT reliability by qualitatively and quantitatively characterizing the memorized samples as well as the model behavior in their vicinity. Based on empirical observations, we develop a simple algorithm which elicits non-memorized translations of memorized samples from the same model, for a large fraction of such samples. Finally, we show that the proposed algorithm could also be leveraged to mitigate memorization in the model through finetuning. We have released the code to reproduce our results at https://github.com/vyraun/Finding-Memo.
翻译:记忆化对若干受限制的自然语言生成(NLG)任务提出了挑战,如神经机器翻译(NMT)等,神经模型对噪音和非典型样品进行记忆化处理,对噪音(网络爬动)数据集产生不利反应;然而,以前对受限制的自然语言生成(NLG)任务的记忆化研究仅侧重于反事实记忆化,将其与幻觉问题联系起来;在这项工作中,我们提出了一种新的廉价算法,用于在受限制的序列生成任务中进行采掘记忆化(在不充分的情况下进行实际培训数据生成),并利用它来研究采掘记忆化及其在NMT的效果。我们证明,采掘记忆化对NMT的可靠性构成严重威胁,办法是定性和定量地定性混集样品以及其周围的模型行为。根据经验观察,我们开发了一种简单的算法,从同一模型中为大量此类样品提取的记忆化样品进行非记忆化翻译(在不充分的情况下进行具体的培训数据生成)。最后,我们表明,拟议的模型还可以通过微调MMemologimal(Wemob)对模型进行升级,从而降低Morizmalimalisal化结果。