Scheduled sampling is an effective method to alleviate the exposure bias problem of neural machine translation. It simulates the inference scene by randomly replacing ground-truth target input tokens with predicted ones during training. Despite its success, its critical schedule strategies are merely based on training steps, ignoring the real-time model competence, which limits its potential performance and convergence speed. To address this issue, we propose confidence-aware scheduled sampling. Specifically, we quantify real-time model competence by the confidence of model predictions, based on which we design fine-grained schedule strategies. In this way, the model is exactly exposed to predicted tokens for high-confidence positions and still ground-truth tokens for low-confidence positions. Moreover, we observe vanilla scheduled sampling suffers from degenerating into the original teacher forcing mode since most predicted tokens are the same as ground-truth tokens. Therefore, under the above confidence-aware strategy, we further expose more noisy tokens (e.g., wordy and incorrect word order) instead of predicted ones for high-confidence token positions. We evaluate our approach on the Transformer and conduct experiments on large-scale WMT 2014 English-German, WMT 2014 English-French, and WMT 2019 Chinese-English. Results show that our approach significantly outperforms the Transformer and vanilla scheduled sampling on both translation quality and convergence speed.
翻译:定时抽样是缓解神经机翻译暴露偏差问题的有效方法,它通过随机替换地面真实目标输入符号,以培训期间预测的符号取代地面真实目标输入符号,来模拟推断场景。尽管取得了成功,但其关键时间表战略仅以培训步骤为基础,忽视了实时模型能力,从而限制了其潜在性能和趋同速度;为了解决这一问题,我们提议采用有自信的定时抽样。具体地,我们用模型预测的可信度来量化实时模型能力,我们据此设计细微的排程战略。这样,模型就完全暴露在高信任位置的预测符号和仍然以低信心位置的地面验证符号上。此外,我们观察香草定时的采样有退化为原始教师强迫模式,因为大多数预测的象征与地面图象一样。因此,根据上述有自信的战略,我们进一步暴露了更吵闹的标志(例如,单词顺序和不正确的单词顺序),而不是高信任标志位置的预测符号。我们评估了2014年在变换式、19年中国质量和2014年大幅变压的英国和WMT(WM)的英国和2014年)的样本上,对英国和中国的大幅的变压的升级和英国的升级的升级。