Non-autoregressive translation (NAT) model achieves a much faster inference speed than the autoregressive translation (AT) model because it can simultaneously predict all tokens during inference. However, its translation quality suffers from degradation compared to AT. And existing NAT methods only focus on improving the NAT model's performance but do not fully utilize it. In this paper, we propose a simple but effective method called "Candidate Soups," which can obtain high-quality translations while maintaining the inference speed of NAT models. Unlike previous approaches that pick the individual result and discard the remainders, Candidate Soups (CDS) can fully use the valuable information in the different candidate translations through model uncertainty. Extensive experiments on two benchmarks (WMT'14 EN-DE and WMT'16 EN-RO) demonstrate the effectiveness and generality of our proposed method, which can significantly improve the translation quality of various base models. More notably, our best variant outperforms the AT model on three translation tasks with 7.6 times speedup.
翻译:与自动递减翻译模型相比,非递减翻译模型的推论速度要快得多,因为它可以同时预测推论期间的所有符号。然而,与AT相比,其翻译质量差强人意。现有的NAT方法只注重改进NAT模型的性能,但没有充分利用它。在本文中,我们提出了一个简单而有效的方法,叫做“Candidate Soups ”,它可以获得高质量的翻译,同时保持NAT模型的推论速度。与以前采用的方法不同,即采集个人结果并丢弃其余部分的方法不同,候选索普(CDS)可以通过模型不确定性充分利用不同候选翻译中的宝贵信息。关于两个基准(WMT'14 EN-DE和WMT'16 EN-RO)的广泛实验表明我们拟议方法的有效性和普遍性,这可以大大改善各种基模型的翻译质量。更值得注意的是,我们的最佳变式比AT模型的三种翻译任务高出7.6倍速度。