We propose DeRenderNet, a deep neural network to decompose the albedo and latent lighting, and render shape-(in)dependent shadings, given a single image of an outdoor urban scene, trained in a self-supervised manner. To achieve this goal, we propose to use the albedo maps extracted from scenes in videogames as direct supervision and pre-compute the normal and shadow prior maps based on the depth maps provided as indirect supervision. Compared with state-of-the-art intrinsic image decomposition methods, DeRenderNet produces shadow-free albedo maps with clean details and an accurate prediction of shadows in the shape-independent shading, which is shown to be effective in re-rendering and improving the accuracy of high-level vision tasks for urban scenes.
翻译:我们提议DeRenderNet,这是一个深层神经网络,用于分解反照和潜伏照明,并制作形状(内)独立阴影,以户外城市场景的单一图像,进行自我监督的培训。为实现这一目标,我们提议使用从视频游戏场景中提取的反光图作为直接监督,并预先根据作为间接监督提供的深度地图绘制正常和影子的先前地图。与最先进的内在图像分解方法相比,DeRenderNet制作了无影反光图,附有干净的细节,准确预测了以外形遮光的阴影。 事实证明,这种图在重塑和提高城市场景高级愿景任务的准确性方面是有效的。