Due to the rise of spherical cameras, monocular 360 depth estimation becomes an important technique for many applications (e.g., autonomous systems). Thus, state-of-the-art frameworks for monocular 360 depth estimation such as bi-projection fusion in BiFuse are proposed. To train such a framework, a large number of panoramas along with the corresponding depth ground truths captured by laser sensors are required, which highly increases the cost of data collection. Moreover, since such a data collection procedure is time-consuming, the scalability of extending these methods to different scenes becomes a challenge. To this end, self-training a network for monocular depth estimation from 360 videos is one way to alleviate this issue. However, there are no existing frameworks that incorporate bi-projection fusion into the self-training scheme, which highly limits the self-supervised performance since bi-projection fusion can leverage information from different projection types. In this paper, we propose BiFuse++ to explore the combination of bi-projection fusion and the self-training scenario. To be specific, we propose a new fusion module and Contrast-Aware Photometric Loss to improve the performance of BiFuse and increase the stability of self-training on real-world videos. We conduct both supervised and self-supervised experiments on benchmark datasets and achieve state-of-the-art performance.
翻译:由于球形照相机的上升,单眼360深度估计成为许多应用(如自主系统)的一个重要技术,因此,提出了单眼360深度估计的最先进框架,如BiFuse的双射聚集。为了培训这样一个框架,需要大量的全景以及激光传感器所捕捉的相应的深度地面真象,这大大增加了数据收集的成本。此外,由于这种数据收集程序耗费时间,将这些方法推广到不同场景的可扩展性成为一项挑战。为此,从360个视频中自我培训一个单眼深度估计网络是缓解这一问题的一种方法。然而,目前没有将双射聚集纳入自我培训计划的现有框架,因为双射聚集能够利用不同投影类型的信息,严重限制了自我监督的性能。在本文件中,我们提议BiFuse++ 探讨双投影集集和自我培训情景的结合。具体而言,我们提议一个新的组合模块和对360个视频的单眼深度估计网络进行单向深度估计,这是缓解这一问题的一种方法。但是,目前还没有将双投影集集集集集成集集集集集集成集成和反校准的自我测试的自我测试,我们提议在稳定性世界的自我测试中将改进了自我测试的自我测试的自我测试的自我测试性能的自我测试性能的自我测试性能。