Multitask learning is a common approach in machine learning, which allows to train multiple objectives with a shared architecture. It has been shown that by training multiple tasks together inference time and compute resources can be saved, while the objectives performance remains on a similar or even higher level. However, in perception related multitask networks only closely related tasks can be found, such as object detection, instance and semantic segmentation or depth estimation. Multitask networks with diverse tasks and their effects with respect to efficiency on one another are not well studied. In this paper we augment the CenterNet anchor-free approach for training multiple diverse perception related tasks together, including the task of object detection and semantic segmentation as well as human pose estimation. We refer to this DNN as Multitask-CenterNet (MCN). Additionally, we study different MCN settings for efficiency. The MCN can perform several tasks at once while maintaining, and in some cases even exceeding, the performance values of its corresponding single task networks. More importantly, the MCN architecture decreases inference time and reduces network size when compared to a composition of single task networks.
翻译:多任务学习是机器学习中的一种常见方法,它能够用一个共同的结构来培训多重目标,已经表明,通过培训多种任务,同时计算时间和计算资源,可以节省,而目标的绩效仍然在类似甚至更高的水平上,然而,在观念上,只能找到与多任务网络密切相关的任务,例如物体探测、实例和语义分解或深度估计。多任务网络具有不同任务及其在相互之间效率方面的影响,没有很好地研究。在本文中,我们加强了中心网络的无锚化方法,以共同培训多种不同认知相关任务,包括对象探测和语义分解的任务以及人造估计。我们把DNN称为多任务网络(Multitask-CenterNet)。此外,我们研究不同的MCN环境效率。MCN可以同时执行几项任务,同时保持并在某些情况下甚至超过其相应的单一任务网络的性能价值。更重要的是,MCN结构会减少推算时间,在与单一任务网络的构成相比,网络规模会缩小。