The Multi-Task Learning (MTL) technique has been widely studied by word-wide researchers. The majority of current MTL studies adopt the hard parameter sharing structure, where hard layers tend to learn general representations over all tasks and specific layers are prone to learn specific representations for each task. Since the specific layers directly follow the hard layers, the MTL model needs to estimate this direct change (from general to specific) as well. To alleviate this problem, we introduce the novel cluster layer, which groups tasks into clusters during training procedures. In a cluster layer, the tasks in the same cluster are further required to share the same network. By this way, the cluster layer produces the general presentation for the same cluster, while produces relatively specific presentations for different clusters. As transitions the cluster layers are used between the hard layers and the specific layers. The MTL model thus learns general representations to specific representations gradually. We evaluate our model with MTL document classification and the results demonstrate the cluster layer is quite efficient in MTL.
翻译:多任务学习技术(MTL)已经由全字研究人员广泛研究,目前的大多数MTL研究采用硬参数共享结构,硬层次往往学习所有任务的一般表述,具体层次容易学习每项任务的具体表述。由于具体层次直接遵循硬层次,MTL模式需要估计这种直接变化(从一般到具体),为了缓解这一问题,我们引入了新颖的集群层,在培训程序期间将任务分组。在一个集群层,同一集群组的任务还需要共享同一网络。这样,集群层为同一集群编制一般表述,同时为不同的集群编制比较具体的表述。随着集群层在硬层次和具体层次之间过渡,集群层在硬层次和具体层次之间被使用。因此,MTL模型需要逐渐了解具体表述的一般表述。我们用MTL文件分类来评估我们的模型,结果显示分组在MTL中的效率很高。