Standard spatial convolutions assume input data with a regular neighborhood structure. Existing methods typically generalize convolution to the irregular point cloud domain by fixing a regular "view" through e.g. a fixed neighborhood size, where the convolution kernel size remains the same for each point. However, since point clouds are not as structured as images, the fixed neighbor number gives an unfortunate inductive bias. We present a novel graph convolution named Difference Graph Convolution (diffConv), which does not rely on a regular view. diffConv operates on spatially-varying and density-dilated neighborhoods, which are further adapted by a learned masked attention mechanism. Experiments show that our model is very robust to the noise, obtaining state-of-the-art performance in 3D shape classification and scene understanding tasks, along with a faster inference speed.
翻译:标准空间共变假设输入数据带有正常的邻居结构。 现有的方法通常是通过固定的邻居规模来固定固定的“ 视野 ”, 通过固定的“ 视野 ”, 通过固定的“ 视野 ”, 通过固定的邻里大小, 凝聚内核的大小与每个点相同。 但是, 由于点云结构不如图像, 固定的邻里编号给出了不幸的感应偏差。 我们展示了名为“ 差异图形共变( diff Conv) ” ( diff Conv) 的新的图形共变形图案。 diff Conv 运行于空间变形和密度拉动的邻里区, 并且通过一个有知识的蒙蔽的注意机制进行进一步调整。 实验显示, 我们的模型非常坚固, 能够适应噪音, 获得3D 形状的状态性能和场景理解任务, 以及更快的发酵速度 。