Current self-training methods such as standard self-training, co-training, tri-training, and others often focus on improving model performance on a single task, utilizing differences in input features, model architectures, and training processes. However, many tasks in natural language processing are about different but related aspects of language, and models trained for one task can be great teachers for other related tasks. In this work, we propose friend-training, a cross-task self-training framework, where models trained to do different tasks are used in an iterative training, pseudo-labeling, and retraining process to help each other for better selection of pseudo-labels. With two dialogue understanding tasks, conversational semantic role labeling and dialogue rewriting, chosen for a case study, we show that the models trained with the friend-training framework achieve the best performance compared to strong baselines.
翻译:目前的各种自我培训方法,如标准自我培训、联合培训、三联培训等,往往侧重于利用投入特点、模式架构和培训过程的差异,改进单一任务的模式绩效。然而,自然语言处理中的许多任务涉及语言的不同但相关的方面,为一项任务而培训的模式可能是从事其他相关任务的优秀教师。在这项工作中,我们建议采用朋友培训、跨任务自我培训框架,在迭代培训、假标签和再培训过程中使用经过培训的不同任务模式,以便相互帮助,更好地选择假标签。通过两项对话理解任务,即谈话性语义角色标签和对话重写,为案例研究所选,我们展示了与强有力的基线相比,经过朋友培训框架培训的模式取得最佳业绩。