This work presents a novel method for predicting vehicle trajectories in highway scenarios using efficient bird's eye view representations and convolutional neural networks. Vehicle positions, motion histories, road configuration, and vehicle interactions are easily included in the prediction model using basic visual representations. The U-net model has been selected as the prediction kernel to generate future visual representations of the scene using an image-to-image regression approach. A method has been implemented to extract vehicle positions from the generated graphical representations to achieve subpixel resolution. The method has been trained and evaluated using the PREVENTION dataset, an on-board sensor dataset. Different network configurations and scene representations have been evaluated. This study found that U-net with 6 depth levels using a linear terminal layer and a Gaussian representation of the vehicles is the best performing configuration. The use of lane markings was found to produce no improvement in prediction performance. The average prediction error is 0.47 and 0.38 meters and the final prediction error is 0.76 and 0.53 meters for longitudinal and lateral coordinates, respectively, for a predicted trajectory length of 2.0 seconds. The prediction error is up to 50% lower compared to the baseline method.
翻译:这项工作是利用高效鸟眼视图显示器和进化神经网络预测高速公路情景中的车辆轨迹的新方法。车辆位置、运动历史、道路配置和车辆互动很容易以基本视觉显示方式纳入预测模型。U-net模型被选为预测内核,以便利用图像到图像回归法产生未来现场的视觉显示。采用了一种方法从生成的图形显示器中提取车辆位置,以实现分像体分辨率。该方法已经使用“防暴”数据集、机载传感器数据集进行了培训和评估。对不同的网络配置和场景展示进行了评估。这项研究发现,使用直线终端层和高斯代表器的6个深度的U-net是最佳的操作配置。使用车道标记无法改善预测性能。平均预测误差为0.47米和0.38米,最后预测误差为0.76米和0.53米。预测误差为2.0秒的预测轨距,预测误差为50米至基线的低。预测误差为50个百分点,比基线低。