Multitask learning (MTL) has recently gained a lot of popularity as a learning paradigm that can lead to improved per-task performance while also using fewer per-task model parameters compared to single task learning. One of the biggest challenges regarding MTL networks involves how to share features across tasks. To address this challenge, we propose the Attentive Task Interaction Network (ATI-Net). ATI-Net employs knowledge distillation of the latent features for each task, then combines the feature maps to provide improved contextualized information to the decoder. This novel approach to introducing knowledge distillation into an attention based multitask network outperforms state of the art MTL baselines such as the standalone MTAN and PAD-Net, with roughly the same number of model parameters.
翻译:多任务学习(MTL)最近作为一种学习范例获得了大量受人欢迎的程度,这种学习模式可以改善每个任务的业绩,同时使用较少的每个任务模式参数,而只使用单一任务学习。MTL网络的最大挑战之一是如何分享不同任务的特点。为了应对这一挑战,我们提议建立“强化任务互动网络(ATI-Net) ” 。 ATI-Net利用对每项任务潜在特征的知识蒸馏,然后将地貌图结合起来,为解码器提供更好的背景信息。这种将知识蒸馏引入关注的多任务网络的新做法,超越了基于关注的多任务网络的艺术MTL基线状态,如独立的MTAN和PAD-Net,其模型参数数量大致相同。