Depth maps are used in a wide range of applications from 3D rendering to 2D image effects such as Bokeh. However, those predicted by single image depth estimation (SIDE) models often fail to capture isolated holes in objects and/or have inaccurate boundary regions. Meanwhile, high-quality masks are much easier to obtain, using commercial auto-masking tools or off-the-shelf methods of segmentation and matting or even by manual editing. Hence, in this paper, we formulate a novel problem of mask-guided depth refinement that utilizes a generic mask to refine the depth prediction of SIDE models. Our framework performs layered refinement and inpainting/outpainting, decomposing the depth map into two separate layers signified by the mask and the inverse mask. As datasets with both depth and mask annotations are scarce, we propose a self-supervised learning scheme that uses arbitrary masks and RGB-D datasets. We empirically show that our method is robust to different types of masks and initial depth predictions, accurately refining depth values in inner and outer mask boundary regions. We further analyze our model with an ablation study and demonstrate results on real applications. More information can be found at https://sooyekim.github.io/MaskDepth/ .
翻译:深度图用于从3D投影到2D图像效应的广泛应用,如Bokeh。然而,通过单一图像深度估计(SIDE)模型预测的深度图往往无法捕捉物体和(或)边界区域不准确的孤立洞穴。与此同时,使用商业自动制模工具或现成的分解和交配方法,甚至人工编辑,高质量的面罩面罩更容易获得。因此,在本文件中,我们提出了一个掩码引导深度改进的新问题,利用通用面罩改进SID模型的深度预测。我们的框架进行层层改进和油漆/油漆,将深度图分解成两个不同的层,由遮罩和反面面罩标注为标志。由于深度和面罩说明的数据集很少,我们建议采用自我监督的学习计划,使用任意的面罩和RGB-D数据集。我们的经验显示,我们的方法对不同种类的面具和初步深度预测十分健全,精确地改进了内外部掩码边界区域的深度值。我们进一步分析我们的深度图案模型,并用一个掩码/Morebismaisk。