Semantic segmentation has been widely investigated in the community, in which the state of the art techniques are based on supervised models. Those models have reported unprecedented performance at the cost of requiring a large set of high quality segmentation masks. To obtain such annotations is highly expensive and time consuming, in particular, in semantic segmentation where pixel-level annotations are required. In this work, we address this problem by proposing a holistic solution framed as a three-stage self-training framework for semi-supervised semantic segmentation. The key idea of our technique is the extraction of the pseudo-masks statistical information to decrease uncertainty in the predicted probability whilst enforcing segmentation consistency in a multi-task fashion. We achieve this through a three-stage solution. Firstly, we train a segmentation network to produce rough pseudo-masks which predicted probability is highly uncertain. Secondly, we then decrease the uncertainty of the pseudo-masks using a multi-task model that enforces consistency whilst exploiting the rich statistical information of the data. We compare our approach with existing methods for semi-supervised semantic segmentation and demonstrate its state-of-the-art performance with extensive experiments.
翻译:在社区广泛调查了语义分解问题,在社区中,艺术技术的状态以监督模型为基础。这些模型报告了前所未有的表现,代价是需要一套高品质分解面罩。要获得这种说明非常昂贵和耗时,特别是在需要像素级注释的语义分解中。在这项工作中,我们提出一个整体解决办法,作为半受监督语义分解的三阶段自我培训框架来解决这一问题。我们技术的关键思想是提取假物质统计信息,以减少预测概率的不确定性,同时以多任务方式执行分解的一致性。我们通过三阶段解决办法实现这一目标。首先,我们培训一个分解网络,以产生粗化的假人形,预测概率非常不确定。第二,我们用一个多任务模型来降低假体的不确定性,该模型在利用丰富的数据统计信息的同时,强制执行一致性。我们将我们的方法与现有的半受监督的语义分解分解方法进行比较,并展示其状态实验。