Road detection is a critically important task for self-driving cars. By employing LiDAR data, recent works have significantly improved the accuracy of road detection. Relying on LiDAR sensors limits the wide application of those methods when only cameras are available. In this paper, we propose a novel road detection approach with RGB being the only input during inference. Specifically, we exploit pseudo-LiDAR using depth estimation, and propose a feature fusion network where RGB and learned depth information are fused for improved road detection. To further optimize the network structure and improve the efficiency of the network. we search for the network structure of the feature fusion module using NAS techniques. Finally, be aware of that generating pseudo-LiDAR from RGB via depth estimation introduces extra computational costs and relies on depth estimation networks, we design a modality distillation strategy and leverage it to further free our network from these extra computational cost and dependencies during inference. The proposed method achieves state-of-the-art performance on two challenging benchmarks, KITTI and R2D.
翻译:使用LiDAR数据,最近的工作大大提高了道路探测的准确性。依靠LiDAR传感器限制这些方法的广泛应用,只要只有摄像头。在本文件中,我们提议采用新的道路探测方法,在推断过程中,只有RGB是唯一的输入物。具体地说,我们利用深度估算来利用假LiDAR,并提议一个功能融合网络,使RGB和所学深层信息结合起来,以便改进道路探测。为了进一步优化网络结构和提高网络效率,我们利用NAS技术搜索地物聚合模块的网络结构。最后,我们意识到通过深度估算从RGB产生伪LiDAR会增加计算成本,并依靠深度估算网络,我们设计一种模式蒸馏战略,利用它进一步使我们的网络摆脱这些额外的计算成本,在推断过程中依赖。拟议方法在两个具有挑战性的基准(KITTI和R2D)上取得了最新业绩。