Changing environments poses a great challenge on the outdoor visual perception and scene understanding for robust long-term autonomous driving and mobile robots, where depth-auxiliary geometric information plays an essential role to the robustness under challenging scenes. Although monocular depth prediction has been well studied recently, there are few work focusing on the depth prediction across multiple environmental conditions, e.g. changing illumination and seasons, owing to the lack of such a real-world dataset and benchmark. In this work, a new cross-season monocular depth prediction dataset SeasonDepth (available on https://seasondepth.github.io) is derived from CMU Visual Localization dataset through structure from motion. To benchmark the depth estimation performance under different environments, we investigate representative and recent state-of-the-art open-source supervised, self-supervised and domain adaptation depth prediction methods from KITTI benchmark using several newly-formulated metrics. Through extensive experimental evaluation on the proposed dataset without fine-tuning, the influence of multiple environments on performance and robustness is analyzed both qualitatively and quantitatively, showing that the long-term monocular depth prediction is far from solved. We further give promising solutions especially with stereo geometry and multi-task sequential self-supervised training to enhance the robustness to changing environments.
翻译:尽管最近对单层深度预测进行了仔细研究,但是由于缺少这种真实世界数据集和基准,很少有工作侧重于在多种环境条件下进行深度预测,例如,由于缺少这种真实世界数据集和基准,因此在变化环境中的环境变化给户外视觉感知和对稳健的长期自主驱动和移动机器人的场景理解带来了巨大挑战。在这项工作中,从CMU视觉本地化数据集(见https://seasson explocation.github.io)中得出新的跨季节单层单层深度预测数据集(见https://seson develop.github.io),在挑战场景中,深度定位信息对稳健性发挥了至关重要的作用。为了对不同环境中的深度估计业绩进行基准评估,我们很少开展侧重于深度预测的工作,例如,由于缺少这种真实世界数据集和基准,不断变化的照明和领域适应深度预测方法。通过对拟议的数据集进行广泛的实验性评估,对多种环境的性能和坚固性进行了分析,从定性和定量分析显示长期单层深度预测是具有希望的,从不断进行自我测测测测的多层环境,从一个更。