Recent breakthroughs in semi-supervised semantic segmentation have been developed through contrastive learning. In prevalent pixel-wise contrastive learning solutions, the model maps pixels to deterministic representations and regularizes them in the latent space. However, there exist inaccurate pseudo-labels which map the ambiguous representations of pixels to the wrong classes due to the limited cognitive ability of the model. In this paper, we define pixel-wise representations from a new perspective of probability theory and propose a Probabilistic Representation Contrastive Learning (PRCL) framework that improves representation quality by taking its probability into consideration. Through modelling the mapping from pixels to representations as the probability via multivariate Gaussian distributions, we can tune the contribution of the ambiguous representations to tolerate the risk of inaccurate pseudo-labels. Furthermore, we define prototypes in the form of distributions, which indicates the confidence of a class, while the point prototype cannot. Moreover, we propose to regularize the distribution variance to enhance the reliability of representations. Taking advantage of these benefits, high-quality feature representations can be derived in the latent space, thereby the performance of semantic segmentation can be further improved. We conduct sufficient experiment to evaluate PRCL on Pascal VOC and CityScapes to demonstrate its superiority. The code is available at https://github.com/Haoyu-Xie/PRCL.
翻译:近些年来,在半监督的语义分解中,通过对比性学习,形成了近些的突破。在流行的像素的对比性学习解决方案中,模型地图像素可以确定在暗地空间的表达方式,并规范这些表达方式。然而,由于模型认知能力有限,存在着不准确的假标签,可以描绘像素与错误类别之间的模棱两可的表达方式。在本文中,我们从概率理论的新角度来定义像素的表达方式,并提出一个概率代表对比学习(PRCL)框架,通过考虑其概率来提高代表性质量。通过模拟从像素到表达方式的表达方式,我们可以通过多变量高斯分布式的概率来模拟这些表达方式的表达方式。我们可以调整模糊的表达方式,以容忍不准确的假标签的风险。此外,我们用分布式的形式界定了模型,表明一个阶级的信心,而点原型则无法。我们提议调整分布差异差异,以提高表达方式的可靠性。我们建议利用这些好处,在暗地空间进一步衍生出高质量的特征表达方式,因此,我们可以在PLS/CLA 进行充分的实验。