Light Detection And Ranging (LiDAR) has been widely used in autonomous vehicles for perception and localization. However, the cost of a high-resolution LiDAR is still prohibitively expensive, while its low-resolution counterpart is much more affordable. Therefore, using low-resolution LiDAR for autonomous driving is an economically viable solution, but the point cloud sparsity makes it extremely challenging. In this paper, we propose a two-stage neural network framework that enables 3D object detection using a low-resolution LiDAR. Taking input from a low-resolution LiDAR point cloud and a monocular camera image, a depth completion network is employed to produce dense point cloud that is subsequently processed by a voxel-based network for 3D object detection. Evaluated with KITTI dataset for 3D object detection in Bird-Eye View (BEV), the experimental result shows that the proposed approach performs significantly better than directly applying the 16-line LiDAR point cloud for object detection. For both easy and moderate cases, our 3D vehicle detection results are close to those using 64-line high-resolution LiDARs.
翻译:光探测和测距(LiDAR)已被广泛用于自主工具,以进行感知和定位。然而,高分辨率激光雷达的成本仍然高得令人望而却步,而低分辨率激光雷达的成本却高得令人望而生畏,而低分辨率激光雷达的成本则更低廉。因此,使用低分辨率激光雷达进行自主驾驶是一个经济上可行的解决方案,但点云雾使得它极具挑战性。在本文中,我们提出了一个两阶段神经网络框架,使3D物体能够使用低分辨率激光雷达进行探测。从低分辨率激光雷达点云和一个单视相机图像中输入,一个深度完成网络用来产生密度点云,随后由基于 voxel 的网络处理,用于3D物体探测。用KITTI 数据集来评估3D物体在Bird-Eye View(BEV)中探测,实验结果表明,拟议的方法比直接应用16线激光雷达点云进行物体探测要好得多。对于简单和中小案例来说,我们的3D车辆探测结果都接近使用64线高分辨率激光雷达。