Domain adaptation allows generative language models to address specific flaws caused by the domain shift of their application. However, the traditional adaptation by further training on in-domain data rapidly weakens the model's ability to generalize to other domains, making the open-ended deployments of the adapted models prone to errors. This work introduces novel training objectives built upon a semantic similarity of the predicted tokens to the reference. Our results show that (1) avoiding the common assumption of a single correct prediction by constructing the training target from tokens' semantic similarity can mitigate catastrophic forgetting during domain adaptation, while (2) preserving the quality of the adaptation, (3) with negligible additions to compute costs. In the broader perspective, the objectives grounded in a soft token alignment pioneer the exploration of the middle ground between the efficient but naive exact-match token-level objectives and expressive but computationally- and resource-intensive sequential objectives.
翻译:域的适应使得基因化语言模型能够解决其应用领域变化造成的具体缺陷,然而,通过对域内数据进行进一步培训进行传统适应,迅速削弱了模型推广到其他领域的能力,使调整后模型的开放式部署容易出现错误。这项工作引入了新的培训目标,其基础是预测符号的语义相似性。我们的结果显示:(1) 通过从象征语的语义相似性构建培训目标,避免一个单一正确预测的共同假设,可以减轻域内适应过程中的灾难性遗忘,(2) 保持适应质量,(3) 与计算成本相加微不足道。从更广的角度来看,以软象征性一致为基础的目标率先探索了高效但天真切的代号级目标和表达但计算和资源密集的相继目标之间的中间地带。