Despite the recent progress in deep learning, most approaches still go for a silo-like solution, focusing on learning each task in isolation: training a separate neural network for each individual task. Many real-world problems, however, call for a multi-modal approach and, therefore, for multi-tasking models. Multi-task learning (MTL) aims to leverage useful information across tasks to improve the generalization capability of a model. This thesis is concerned with multi-task learning in the context of computer vision. First, we review existing approaches for MTL. Next, we propose several methods that tackle important aspects of multi-task learning. The proposed methods are evaluated on various benchmarks. The results show several advances in the state-of-the-art of multi-task learning. Finally, we discuss several possibilities for future work.
翻译:尽管在深层学习方面最近取得了进展,但大多数方法仍倾向于一个筒仓式的解决方案,侧重于孤立地学习每项任务:为每项任务培训单独的神经网络。然而,许多现实世界问题要求采用多模式方法,因此要求采用多任务模式。多任务学习(MTL)旨在利用跨任务的有用信息,提高模式的普及能力。这一论文涉及计算机愿景背景下的多任务学习。首先,我们审查现有的多任务学习方法。接下来,我们提出了处理多任务学习重要方面的若干方法。对拟议方法进行了各种基准评估。结果显示,在多任务学习的最先进技术方面取得了一些进展。最后,我们讨论了未来工作的若干可能性。