Neural volume rendering enables photo-realistic renderings of a human performer in free-view, a critical task in immersive VR/AR applications. But the practice is severely limited by high computational costs in the rendering process. To solve this problem, we propose the UV Volumes, a new approach that can render an editable free-view video of a human performer in realtime. It separates the high-frequency (i.e., non-smooth) human appearance from the 3D volume, and encodes them into 2D neural texture stacks (NTS). The smooth UV volumes allow much smaller and shallower neural networks to obtain densities and texture coordinates in 3D while capturing detailed appearance in 2D NTS. For editability, the mapping between the parameterized human model and the smooth texture coordinates allows us a better generalization on novel poses and shapes. Furthermore, the use of NTS enables interesting applications, e.g., retexturing. Extensive experiments on CMU Panoptic, ZJU Mocap, and H36M datasets show that our model can render 960 * 540 images in 30FPS on average with comparable photo-realism to state-of-the-art methods. The project and supplementary materials are available at https://github.com/fanegg/UV-Volumes.
翻译:为了解决这个问题,我们建议了UV卷,这是一个可以实时编辑人类表演者可自由视觉视频的新办法。它可以将高频(即非摩擦)人类外观与3D卷分开,并将其编码为2D神经质质素堆(NTS),光滑的UV卷允许使用更小、更浅的神经网络获取3D的密度和纹理坐标,同时在2D NTS中捕捉详细的外观。对于可编辑性,在参数化人类模型和光滑的纹理坐标之间绘制地图,使我们能够更好地对新的外观和形状进行概括。此外,使用NTS系统可以进行有趣的应用,例如,可以重新输入。在CMU60光学、ZJU Mocap和H36MM-Restreal图像库进行广泛的实验,在5DMU40光学、ZJU Mocap和H36-MM图像库中,可以使我们的模型和30S-FPS的模型和模型中可比较的30S-FPA图像材料显示9的可比较的模型和30S-FPS-S-S-S-FPS-S-S-S-S-S-S-S-S-S-S-Syalalalalals 的可比较的模型和Syalalalals 的模型的可比较的模型和S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-Syalsalalalalalalalal 的模型的模型的模型的可比较性图像材料。