We present PersonNeRF, a method that takes a collection of photos of a subject (e.g. Roger Federer) captured across multiple years with arbitrary body poses and appearances, and enables rendering the subject with arbitrary novel combinations of viewpoint, body pose, and appearance. PersonNeRF builds a customized neural volumetric 3D model of the subject that is able to render an entire space spanned by camera viewpoint, body pose, and appearance. A central challenge in this task is dealing with sparse observations; a given body pose is likely only observed by a single viewpoint with a single appearance, and a given appearance is only observed under a handful of different body poses. We address this issue by recovering a canonical T-pose neural volumetric representation of the subject that allows for changing appearance across different observations, but uses a shared pose-dependent motion field across all observations. We demonstrate that this approach, along with regularization of the recovered volumetric geometry to encourage smoothness, is able to recover a model that renders compelling images from novel combinations of viewpoint, pose, and appearance from these challenging unstructured photo collections, outperforming prior work for free-viewpoint human rendering.
翻译:我们提出人信息,这是一个收集多年来以任意的外形和外形拍摄的某一主题(如罗杰·费德勒)照片的方法,它收集了多年来以任意的外形和外形拍摄的某一主题(如罗杰·费德勒)的照片,使该主题具有任意的新颖组合观点、身体外形和外观。 人信息信息,它构建了该主题的定制神经体积3D模型,能够使整个空间以相机视角、身体外观和外观覆盖。 这项任务中的一项中心挑战涉及零散的观测; 特定身体的外形可能只通过单一的视角观察到,而特定外观只出现在少数不同的身体外观下。 我们通过恢复该主题的罐形外观神经神经体积代表来解决这一问题,允许在不同观测中改变外观的外观,但在所有观测中都使用一个共同的外观运动场。 我们证明,这一方法,加上对已回收的体积几率测量法度进行规范,能够恢复一个模型,从这些富有挑战性的观点、外观和外观的新组合中产生令人信服的图像。