Since the preparation of labeled data for training semantic segmentation networks of point clouds is a time-consuming process, weakly supervised approaches have been introduced to learn from only a small fraction of data. These methods are typically based on learning with contrastive losses while automatically deriving per-point pseudo-labels from a sparse set of user-annotated labels. In this paper, our key observation is that the selection of what samples to annotate is as important as how these samples are used for training. Thus, we introduce a method for weakly supervised segmentation of 3D scenes that combines self-training with active learning. The active learning selects points for annotation that likely result in performance improvements to the trained model, while the self-training makes efficient use of the user-provided labels for learning the model. We demonstrate that our approach leads to an effective method that provides improvements in scene segmentation over previous works and baselines, while requiring only a small number of user annotations.
翻译:由于为培训点云的语义分解网络编制有标签的数据是一个耗时的过程,因此采用了监督不力的方法从一小部分数据中学习。这些方法通常以学习和对比性损失为基础,而自动从稀少的一组用户附加说明的标签中产生每个点的假标签。在本文中,我们的主要观察意见是,选择哪些样本作为注释与这些样本如何用于培训同样重要。因此,我们引入了一种将自我培训和积极学习相结合的3D场景进行监督不力的分解的方法。积极学习的评分点可能导致改进培训模式的性能,而自我培训则是有效利用用户提供的标签来学习模型。我们证明,我们的方法导致一种有效的方法,为先前的工作和基线提供现场分解方面的改进,同时只需要少量的用户说明。