Recently, 3D lane detection has been an actively developing area in autonomous driving which is the key to routing the vehicle. However, the previous work did not balance performance and effectiveness.This work proposes a deployment-oriented monocular 3D lane detector with only naive CNN and FC layers. This detector achieved state-of-the-art results on the Apollo 3D Lane Synthetic dataset and OpenLane real-world dataset with 96 FPS runtime speed. We conduct three techniques in our detector: (1) Virtual Camera eliminates the difference in poses of cameras mounted on different vehicles. (2) Spatial Transformation Pyramid as a light-weighed front-view to bird-eye view transformer can utilize multiscale image-view featmaps. (3) YOLO-Style Representation makes a good balance between bird-eye view resolution and runtime speed, and it can reduce the inefficiency caused by the class imbalance due to the sparsity of the lane detection task during training. Experimental results show that our work outperforms state-of-the-art approaches by 10.6% F1-Score on OpenLane dataset and 4.0% F1-Score on Apollo 3D synthetic dataset and with speed of 96 FPS. The source code will release at https://github.com/hm-gigo-team/bev_lane_det.
翻译:最近,3D车道探测是自动驾驶中一个积极开发的领域,这是车辆航向的关键。然而,先前的工作没有平衡性能和效果。本项工作提议仅使用天真的CNN和FC层进行面向部署的单视3D车道探测仪。该探测器在阿波罗3D车道合成合成数据集和OpenLane现实世界数据集上取得了最先进的结果,达到96 FPS运行速度。我们在探测器中采用了三种技术:(1)虚拟相机消除了不同车辆上安装的相机的配置差异。(2)空间变形金字机作为鸟眼视图的轻比邻前视前视变形器可以使用多级图像视图faatmaps。(3) YOLO-Style代表在Applo 3D视解析和运行速度之间保持了良好的平衡,它可以降低班级失衡造成的效率,原因是在培训期间对车道探测任务过于紧张。实验结果显示,我们的工作超越了艺术状态,在10.6% F1-Scream-Dream-S-Sqoal数据集上将使用96-FlicomalSet数据集的10/Frevream-Flistram-D/Flistram/Frevreax10。