Fine-tuning is known to improve NLP models by adapting an initial model trained on more plentiful but less domain-salient examples to data in a target domain. Such domain adaptation is typically done using one stage of fine-tuning. We demonstrate that gradually fine-tuning in a multi-stage process can yield substantial further gains and can be applied without modifying the model or learning objective.
翻译:据了解,微调通过调整一个初步模型,根据目标领域的数据对经过更丰富但更不具有领域针对性的实例培训的初始模型进行修改,从而改进了NLP模型。这类领域调整通常使用一个微调阶段进行。我们证明,在多阶段进程中逐步微调可以产生更大的进一步收益,并且可以在不修改模型或学习目标的情况下加以应用。