Deep learning approaches have achieved great success in the field of Natural Language Processing (NLP). However, deep neural models often suffer from overfitting and data scarcity problems that are pervasive in NLP tasks. In recent years, Multi-Task Learning (MTL), which can leverage useful information of related tasks to achieve simultaneous performance improvement on multiple related tasks, has been used to handle these problems. In this paper, we give an overview of the use of MTL in NLP tasks. We first review MTL architectures used in NLP tasks and categorize them into four classes, including the parallel architecture, hierarchical architecture, modular architecture, and generative adversarial architecture. Then we present optimization techniques on loss construction, data sampling, and task scheduling to properly train a multi-task model. After presenting applications of MTL in a variety of NLP tasks, we introduce some benchmark datasets. Finally, we make a conclusion and discuss several possible research directions in this field.
翻译:深入的学习方法在自然语言处理领域取得了巨大成功,然而,深神经模型往往因过度装配和数据短缺问题而成为国家语言处理任务中普遍存在的问题。近年来,多任务学习(MTL)能够利用相关任务方面的有用信息,同时改进多重相关任务的业绩,用于处理这些问题。本文概述了在自然语言处理任务中使用MTL的情况。我们首先审查国家语言处理任务中使用的MTL结构,并将其分为四类,包括平行结构、等级结构、模块架构和基因化对抗结构。然后,我们介绍损失建造、数据抽样和任务时间安排的最优化技术,以适当培训一个多任务模式。在介绍MTL在各种国家语言处理任务中的应用情况之后,我们介绍一些基准数据集。最后,我们得出结论,并讨论该领域的若干可能的研究方向。