Autonomous driving has been an active area of research and development, with various strategies being explored for decision-making in autonomous vehicles. Rule-based systems, decision trees, Markov decision processes, and Bayesian networks have been some of the popular methods used to tackle the complexities of traffic conditions and avoid collisions. However, with the emergence of deep learning, many researchers have turned towards CNN-based methods to improve the performance of collision avoidance. Despite the promising results achieved by some CNN-based methods, the failure to establish correlations between sequential images often leads to more collisions. In this paper, we propose a CNN-based method that overcomes the limitation by establishing feature correlations between regions in sequential images using variants of attention. Our method combines the advantages of CNN in capturing regional features with a bi-directional LSTM to enhance the relationship between different local areas. Additionally, we use an encoder to improve computational efficiency. Our method takes "Bird's Eye View" graphs generated from camera and LiDAR sensors as input, simulates the position (x, y) and head offset angle (Yaw) to generate future trajectories. Experiment results demonstrate that our proposed method outperforms existing vision-based strategies, achieving an average of only 3.7 collisions per 1000 miles of driving distance on the L5kit test set. This significantly improves the success rate of collision avoidance and provides a promising solution for autonomous driving.
翻译:自主驱动是研发的一个积极领域,探索了自主车辆决策的各种战略。基于规则的系统、决策树、Markov决策程序和Bayesian网络是用来应对交通状况复杂和避免碰撞的一些流行方法。然而,随着深层次的学习的出现,许多研究人员转向以CNN为基础的方法来提高避免碰撞的性能。尽管有线电视新闻网的一些方法取得了有希望的结果,但未能在相继图像之间建立关联性往往导致更多的碰撞。在本文件中,我们提议一种基于CNN的方法,通过利用关注变量在相继图像中建立不同区域之间的特征相关性来克服限制。我们的方法结合CNN在用双向LSTM获取区域特征以加强不同地方之间关系的优势。此外,我们使用一个编码器来提高计算性效率。我们的方法采用“Bird's眼视图”图作为投入,模拟了基于CDAR传感器的位置(x,y)和头顶偏角(Yaw),以生成一种具有希望的距离的图像率的图像率测试结果,即:我们目前平均的图像率测试结果,只能提供一种稳定的图像率。</s>