Multi-task learning, in which several tasks are jointly learned by a single model, allows NLP models to share information from multiple annotations and may facilitate better predictions when the tasks are inter-related. This technique, however, requires annotating the same text with multiple annotation schemes which may be costly and laborious. Active learning (AL) has been demonstrated to optimize annotation processes by iteratively selecting unlabeled examples whose annotation is most valuable for the NLP model. Yet, multi-task active learning (MT-AL) has not been applied to state-of-the-art pre-trained Transformer-based NLP models. This paper aims to close this gap. We explore various multi-task selection criteria in three realistic multi-task scenarios, reflecting different relations between the participating tasks, and demonstrate the effectiveness of multi-task compared to single-task selection. Our results suggest that MT-AL can be effectively used in order to minimize annotation efforts for multi-task NLP models.
翻译:多任务学习,通过单一模式共同学习若干任务,使NLP模式能够分享多份说明的信息,并有利于在任务相互关联时作出更好的预测。然而,这一技术要求用费用昂贵和费力的多种批注计划来说明相同的文本。积极学习(AL)通过迭代选择对NLP模式最有价值的未贴标签的例子来优化批注过程。然而,多任务积极学习(MT-AL)尚未应用到最先进的预先培训的基于NLP模式。本文旨在缩小这一差距。我们在三种现实的多任务设想中探索各种多任务选择标准,反映参与任务之间的不同关系,并展示多任务与单任务选择的实效。我们的结果表明,MT-AL可以有效地用于尽量减少多任务NLP模式的批注努力。