Vehicle velocity and inter-vehicle distance estimation are essential for ADAS (Advanced driver-assistance systems) and autonomous vehicles. To save the cost of expensive ranging sensors, recent studies focus on using a low-cost monocular camera to perceive the environment around the vehicle in a data-driven fashion. Existing approaches treat each vehicle independently for perception and cause inconsistent estimation. Furthermore, important information like context and spatial relation in 2D object detection is often neglected in the velocity estimation pipeline. In this paper, we explore the relationship between vehicles of the same frame with a global-relative-constraint (GLC) loss to encourage consistent estimation. A novel multi-stream attention network (MSANet) is proposed to extract different aspects of features, e.g., spatial and contextual features, for joint vehicle velocity and inter-vehicle distance estimation. Experiments show the effectiveness and robustness of our proposed approach. MSANet outperforms state-of-the-art algorithms on both the KITTI dataset and TuSimple velocity dataset.
翻译:车辆速度和车辆间距离估计对于ADAS(高级助运系统)和自主车辆至关重要。为了节省昂贵的测距传感器的费用,最近的研究侧重于使用低成本单筒照相机以数据驱动的方式观察车辆周围的环境。现有办法对每部车辆分别进行观察,并造成不一致的估计。此外,在速度估计管道中往往忽视了2D物体探测中的背景和空间关系等重要信息。在本文件中,我们探讨了同一框架的车辆与全球调控限制损失之间的关系,以鼓励一致估计。建议建立一个新的多流注意网络,以提取不同特点的不同方面,例如空间和背景特征,用于联合车辆速度和车辆间距离估计。实验表明我们拟议办法的有效性和稳健性。AMSNet在KITTI数据集和Tusoima速度数据集上都比最新算法。