Autonomous vehicles increasingly rely on cameras to provide the input for perception and scene understanding and the ability of these models to classify their environment and objects, under adverse conditions and image noise is crucial. When the input is, either unintentionally or through targeted attacks, deteriorated, the reliability of autonomous vehicle is compromised. In order to mitigate such phenomena, we propose DriveGuard, a lightweight spatio-temporal autoencoder, as a solution to robustify the image segmentation process for autonomous vehicles. By first processing camera images with DriveGuard, we offer a more universal solution than having to re-train each perception model with noisy input. We explore the space of different autoencoder architectures and evaluate them on a diverse dataset created with real and synthetic images demonstrating that by exploiting spatio-temporal information combined with multi-component loss we significantly increase robustness against adverse image effects reaching within 5-6% of that of the original model on clean images.
翻译:自主车辆越来越依赖相机提供感知和场景理解信息,这些模型在不利条件和图像噪音下对其环境和物体进行分类的能力也越来越重要。当这种输入无意中或通过定向袭击而恶化时,自主车辆的可靠性就会受损。为了缓解这些现象,我们提议DriveGuard,一个轻量的时空自动自动编码器,作为强化自主车辆图像分割过程的一种解决办法。首先用驱动器Guard处理相机图像,我们提供了一个比用噪音输入对每个感知模型进行再培训更加普遍的解决方案。我们探索不同自动编码器结构的空间,并用以真实和合成图像制作的多种数据集对其进行评价,表明通过利用波形时空信息,再加上多构件损失,我们大大提高了抗不良图像影响的能力,使其达到原始图像模型中对清洁图像的5-6%。