Curriculum learning has shown promising improvements in multiple domains by training machine learning models from easy samples to hard ones. Previous works which either design rules or train models for scoring the difficulty highly rely on task-specific expertise, and cannot generalize. Inspired by the ``easy-to-hard'' intuition, we propose to do in-sample curriculum learning for natural language generation tasks. Our learning strategy starts training the model to generate the last few words, i.e., do sequence completion, and gradually extends to generate the whole output sequence. Comprehensive experiments show that it generalizes well to different tasks and achieves significant improvements over strong baselines.
翻译:课程学习通过培训从简单样本到硬样本的机器学习模式,在多个领域显示出了有希望的改善。 以往设计规则或培训模型以克服困难的工作高度依赖特定任务的专门知识,无法概括。 在“ 简单到硬”的直觉的启发下,我们提议为自然语言生成任务开展类比课程学习。 我们的学习战略开始培训模型,以生成最后几个词,即完成序列,并逐步扩展以生成整个产出序列。 全面实验显示,它非常概括不同的任务,在强大的基线上实现显著改进。