Previous face inverse rendering methods often require synthetic data with ground truth and/or professional equipment like a lighting stage. However, a model trained on synthetic data or using pre-defined lighting priors is typically unable to generalize well for real-world situations, due to the gap between synthetic data/lighting priors and real data. Furthermore, for common users, the professional equipment and skill make the task expensive and complex. In this paper, we propose a deep learning framework to disentangle face images in the wild into their corresponding albedo, normal, and lighting components. Specifically, a decomposition network is built with a hierarchical subdivision strategy, which takes image pairs captured from arbitrary viewpoints as input. In this way, our approach can greatly mitigate the pressure from data preparation, and significantly broaden the applicability of face inverse rendering. Extensive experiments are conducted to demonstrate the efficacy of our design, and show its superior performance in face relighting over other state-of-the-art alternatives. {Our code is available at \url{https://github.com/AutoHDR/HD-Net.git}}
翻译:先前的反面转换方法往往需要具有地面真象和/或像照明阶段这样的专业设备的合成数据,然而,由于合成数据/亮亮前程与真实数据之间的差距,经过合成数据或使用预先界定的照明前程的模型通常无法在现实世界中很好地推广。此外,对于普通用户来说,专业设备和技能使得任务既昂贵又复杂。在本文件中,我们提议了一个深层次学习框架,将野生图像与相应的反光、正常和照明组件分解开来。具体地说,一个分解网络是用一个等级分层战略建成的分层网络,从任意的观点中捕捉到的图像配对,作为投入。这样,我们的方法可以大大减轻数据编制的压力,大大扩大面部的可反面应用性。进行了广泛的实验,以展示我们设计的功效,并显示其优异面亮度,以示其他状态的替代方法。 {OUR代码可在以下s://github.com/AutHDR/HD-Net_}