Multi-task learning has recently become a promising solution for a comprehensive understanding of complex scenes. Not only being memory-efficient, multi-task models with an appropriate design can favor exchange of complementary signals across tasks. In this work, we jointly address 2D semantic segmentation, and two geometry-related tasks, namely dense depth, surface normal estimation as well as edge estimation showing their benefit on indoor and outdoor datasets. We propose a novel multi-task learning architecture that exploits pair-wise cross-task exchange through correlation-guided attention and self-attention to enhance the average representation learning for all tasks. We conduct extensive experiments considering three multi-task setups, showing the benefit of our proposal in comparison to competitive baselines in both synthetic and real benchmarks. We also extend our method to the novel multi-task unsupervised domain adaptation setting. Our code is available at https://github.com/cv-rits/DenseMTL.
翻译:多任务学习最近已成为全面了解复杂场景的有希望的解决办法。不仅是记忆效率高、设计适当的多任务模型,还有利于在各项任务之间交换互补信号。在这项工作中,我们共同处理2D语义分割和两个与几何有关的任务,即密集深度、表面正常估计和边缘估计,表明其在室内和室外数据集上的益处。我们提议了一个新的多任务学习架构,通过相关引导的关注和自我关注,利用对等跨任务交流,加强所有任务的平均代表性学习。我们进行了广泛的实验,考虑了三个多任务设置,展示了我们提案在综合和实际基准中与竞争性基线相比的益处。我们还将我们的方法扩大到新的多任务且不受监督的域适应设置。我们的代码可在https://github.com/cv-rits/DenseMTL查阅。