Several popular computer vision (CV) datasets, specifically employed for Object Detection (OD) in autonomous driving tasks exhibit biases due to a range of factors including weather and lighting conditions. These biases may impair a model's generalizability, rendering it ineffective for OD in novel and unseen datasets. Especially, in autonomous driving, it may prove extremely high risk and unsafe for the vehicle and its surroundings. This work focuses on understanding these datasets better by identifying such "good-weather" bias. Methods to mitigate such bias which allows the OD models to perform better and improve the robustness are also demonstrated. A simple yet effective OD framework for studying bias mitigation is proposed. Using this framework, the performance on popular datasets is analyzed and a significant difference in model performance is observed. Additionally, a knowledge transfer technique and a synthetic image corruption technique are proposed to mitigate the identified bias. Finally, using the DAWN dataset, the findings are validated on the OD task, demonstrating the effectiveness of our techniques in mitigating real-world "good-weather" bias. The experiments show that the proposed techniques outperform baseline methods by averaged fourfold improvement.
翻译:专门用于自动驾驶任务中的物体探测(OD)的几种流行计算机视觉(CV)数据集由于天气和照明条件等一系列因素而表现出偏差。这些偏差可能会损害模型的通用性,使其在新颖和看不见的数据集中对OD无效。特别是,在自动驾驶方面,它可能证明风险极高,对车辆及其周围环境来说不安全。这项工作的重点是通过查明这种“良好天气”的偏差来更好地理解这些数据集。还展示了减少这种偏差的方法,这种偏差使OD模型能够更好地发挥作用并改进稳健性。提议了一个简单而有效的研究减少偏差的OD框架。利用这个框架,对大众数据集的性能进行了分析,并观察到模型性能的重大差异。此外,还提出了知识转让技术和合成图像腐败技术,以缓解已查明的偏差。最后,利用DAWN数据集,对OD任务的研究结果进行了验证,表明我们在减轻现实世界“良好天气”偏差方面的技术的有效性。实验表明,拟议的技术以平均四倍的改进方式超越了基线方法。