We present a novel and flexible architecture for point cloud segmentation with dual-representation iterative learning. In point cloud processing, different representations have their own pros and cons. Thus, finding suitable ways to represent point cloud data structure while keeping its own internal physical property such as permutation and scale-invariant is a fundamental problem. Therefore, we propose our work, DRINet, which serves as the basic network structure for dual-representation learning with great flexibility at feature transferring and less computation cost, especially for large-scale point clouds. DRINet mainly consists of two modules called Sparse Point-Voxel Feature Extraction and Sparse Voxel-Point Feature Extraction. By utilizing these two modules iteratively, features can be propagated between two different representations. We further propose a novel multi-scale pooling layer for pointwise locality learning to improve context information propagation. Our network achieves state-of-the-art results for point cloud classification and segmentation tasks on several datasets while maintaining high runtime efficiency. For large-scale outdoor scenarios, our method outperforms state-of-the-art methods with a real-time inference speed of 62ms per frame.
翻译:我们提出了一个具有双重代表性的迭代学习的点云分割新而灵活的结构。 在点云处理中,不同代表有各自的利弊。 因此,找到适当的方法代表点云数据结构,同时保留其自身的内部物理属性,如变换和比例变化等,这是一个根本性问题。 因此,我们建议我们的工作,DRINet,它是双重代表性学习的基本网络结构,在特征传输和较低计算成本方面具有极大的灵活性,特别是对于大型点云。 DRINet主要由两个模块组成,称为 Sprass Porm-Voxel Fatural Explicationon 和 Sprass Voxel-Point Explicationon。通过迭代这两个模块,可以在两个不同的代表之间传播特征。我们进一步提议一个新的多尺度集合层,用于点化地点学习,以改进背景信息传播。我们的网络在保持高运行效率的同时,在几个数据集的点云分类和分解任务方面实现了最先进的结果。 对于大型室外景情景,我们的方法在实时速度62米的每个框架上超越了状态。