We present an algorithm for re-rendering a person from a single image under arbitrary poses. Existing methods often have difficulties in hallucinating occluded contents photo-realistically while preserving the identity and fine details in the source image. We first learn to inpaint the correspondence field between the body surface texture and the source image with a human body symmetry prior. The inpainted correspondence field allows us to transfer/warp local features extracted from the source to the target view even under large pose changes. Directly mapping the warped local features to an RGB image using a simple CNN decoder often leads to visible artifacts. Thus, we extend the StyleGAN generator so that it takes pose as input (for controlling poses) and introduces a spatially varying modulation for the latent space using the warped local features (for controlling appearances). We show that our method compares favorably against the state-of-the-art algorithms in both quantitative evaluation and visual comparison.
翻译:现有方法往往难以在保存源图像的身份和细细细节的同时,对隐蔽内容产生真实的幻觉。 我们首先学会用人体对称性来对身体表面纹理和源图像之间的对应领域进行油漆。 被涂漆的通信领域使我们能够将从源中提取的本地特征(即使是在较大范围内)转移到目标视图中。 直接用简单的CNN解码器将扭曲的本地特征映射成 RGB 图像,常常导致可见的文物。 因此,我们扩展StyleGAN 生成器,以便它作为输入(控制外形),并使用扭曲的本地特征(控制外观)对潜在空间进行空间进行空间不同的空间调整。 我们显示,在定量评估和视觉比较中,我们的方法优于最先进的算法。