Accurate and robust trajectory predictions of road users are needed to enable safe automated driving. To do this, machine learning models are often used, which can show erratic behavior when presented with previously unseen inputs. In this work, two environment-aware models (MotionCNN and MultiPath++) and two common baselines (Constant Velocity and an LSTM) are benchmarked for robustness against various perturbations that simulate functional insufficiencies observed during model deployment in a vehicle: unavailability of road information, late detections, and noise. Results show significant performance degradation under the presence of these perturbations, with errors increasing up to +1444.8\% in commonly used trajectory prediction evaluation metrics. Training the models with similar perturbations effectively reduces performance degradation, with error increases of up to +87.5\%. We argue that despite being an effective mitigation strategy, data augmentation through perturbations during training does not guarantee robustness towards unforeseen perturbations, since identification of all possible on-road complications is unfeasible. Furthermore, degrading the inputs sometimes leads to more accurate predictions, suggesting that the models are unable to learn the true relationships between the different elements in the data.
翻译:摘要: 为了实现安全的自动驾驶,需要准确和鲁棒的路用户轨迹预测。为此,通常使用机器学习模型,但这些模型在遇到以前从未见过的输入时可能会出现奇怪的行为。在本文中,我们对两个环境感知模型(MotionCNN和MultiPath++)和两个常见的基准模型(Constant Velocity和LSTM)进行了基准测试,以评估其针对模型部署过程中观察到的各种功能缺陷的鲁棒性:道路信息不可用、延迟检测和噪声。结果表明,在存在这些扰动的情况下,性能会显著下降,常用的轨迹预测评估指标的误差增加高达+1444.8%。通过类似的扰动训练模型可以有效地降低性能下降,误差最高增加了+87.5%。我们认为,尽管是一种有效的缓解策略,但通过训练的数据扩充并不能保证针对意外扰动的鲁棒性,因为识别所有可能出现的路上问题是不可行的。此外,降低输入有时会导致更准确的预测,这表明模型无法学习数据中不同元素之间的真实关系。