Depth map super-resolution (DSR) has been a fundamental task for 3D computer vision. While arbitrary scale DSR is a more realistic setting in this scenario, previous approaches predominantly suffer from the issue of inefficient real-numbered scale upsampling. To explicitly address this issue, we propose a novel continuous depth representation for DSR. The heart of this representation is our proposed Geometric Spatial Aggregator (GSA), which exploits a distance field modulated by arbitrarily upsampled target gridding, through which the geometric information is explicitly introduced into feature aggregation and target generation. Furthermore, bricking with GSA, we present a transformer-style backbone named GeoDSR, which possesses a principled way to construct the functional mapping between local coordinates and the high-resolution output results, empowering our model with the advantage of arbitrary shape transformation ready to help diverse zooming demand. Extensive experimental results on standard depth map benchmarks, e.g., NYU v2, have demonstrated that the proposed framework achieves significant restoration gain in arbitrary scale depth map super-resolution compared with the prior art. Our codes are available at https://github.com/nana01219/GeoDSR.
翻译:3D 计算机视野的一项基本任务就是深度图超分辨率(DSR) 。 虽然任意比例的DSR是这一假设情景中更现实的设置,但以往的做法主要受到低效率实际数量的规模扩大抽样问题的影响。为了明确解决这一问题,我们提议为DSR提供新的连续深度代表。这一表述的核心是我们提议的几何空间聚合器(GSA),它利用了任意上标的目标网格来调节的距离场,通过这个网格,将几何信息明确引入特征聚合和目标生成中。此外,我们用GeoDSR(GeoDSR)作为GeoDSR(GeoDSR)的砖块,它拥有一种原则性方法,可以构建地方坐标和高分辨率输出结果之间的功能映射,赋予我们的模型以任意形状转换的优势,可以帮助多样化的缩放需求。关于标准深度地图基准的广泛实验结果,例如NYU v2,表明拟议框架在任意规模深度地图超级分辨率与先前的艺术相比取得了重大恢复收益。我们的代码可在 https://githhubub.com/nan1101.Geo。