LiDAR-based 3D object detection is an important task for autonomous driving and current approaches suffer from sparse and partial point clouds of distant and occluded objects. In this paper, we propose a novel two-stage approach, namely PC-RGNN, dealing with such challenges by two specific solutions. On the one hand, we introduce a point cloud completion module to recover high-quality proposals of dense points and entire views with original structures preserved. On the other hand, a graph neural network module is designed, which comprehensively captures relations among points through a local-global attention mechanism as well as multi-scale graph based context aggregation, substantially strengthening encoded features. Extensive experiments on the KITTI benchmark show that the proposed approach outperforms the previous state-of-the-art baselines by remarkable margins, highlighting its effectiveness.
翻译:在本文件中,我们提出了一个新的两阶段办法,即PC-RGNN,通过两个具体办法应对此类挑战。一方面,我们引入了一个点云完成模块,以恢复由密集点组成的高质量建议和保留原有结构的完整观点。另一方面,设计了一个图形神经网络模块,通过地方-全球关注机制和多尺度图表背景汇总全面反映各点之间的关系,大大加强了编码特征。关于KITTI基准的广泛实验表明,拟议的方法以显著的边缘优势超越了以前的最先进的基线,突出了其有效性。