Current language generation models suffer from issues such as repetition, incoherence, and hallucinations. An often-repeated hypothesis is that this brittleness of generation models is caused by the training and the generation procedure mismatch, also referred to as exposure bias. In this paper, we verify this hypothesis by analyzing exposure bias from an imitation learning perspective. We show that exposure bias leads to an accumulation of errors, analyze why perplexity fails to capture this accumulation, and empirically show that this accumulation results in poor generation quality. Source code to reproduce these experiments is available at https://github.com/kushalarora/quantifying_exposure_bias
翻译:当代语言生成模型存在重复、不一致和幻觉等问题。 经常重复的假设是,这种生成模型的萎缩是由培训和生成程序不匹配造成的,也被称为暴露偏差。 在本文中,我们通过从模仿学习角度分析暴露偏差来验证这一假设。 我们表明,暴露偏差会导致错误的积累,分析为什么不理解未能捕捉这种积累,并用经验证明这种积累导致生成质量差。 复制这些实验的源代码可在 https://github.com/kushalarora/quantization_Explosure_bias查阅。