Recent neural rendering approaches greatly improve image quality, reaching near photorealism. However, the underlying neural networks have high runtime, precluding telepresence and virtual reality applications that require high resolution at low latency. The sequential dependency of layers in deep networks makes their optimization difficult. We break this dependency by caching information from the previous frame to speed up the processing of the current one with an implicit warp. The warping with a shallow network reduces latency and the caching operations can further be parallelized to improve the frame rate. In contrast to existing temporal neural networks, ours is tailored for the task of rendering novel views of faces by conditioning on the change of the underlying surface mesh. We test the approach on view-dependent rendering of 3D portrait avatars, as needed for telepresence, on established benchmark sequences. Warping reduces latency by 70$\%$ (from 49.4ms to 14.9ms on commodity GPUs) and scales frame rates accordingly over multiple GPUs while reducing image quality by only 1$\%$, making it suitable as part of end-to-end view-dependent 3D teleconferencing applications. Our project page can be found at: https://yu-frank.github.io/lowlatency/.
翻译:最近的神经转换方法大大提高了图像质量,接近光现实。然而,根基神经网络的运行时间长,排除了远程存在和虚拟现实应用,这些都要求低悬浮度高分辨率的高分辨率。深网络层的相继依赖性使其难以优化。我们打破了这一依赖性,将信息从前一个框架封存,以加快当前网络的暗曲处理速度。使用浅网络的扭曲可以进一步降低延缓度,缓冲操作可以进一步与提高框架率平行。与现有的时空神经网络不同,我们的设计是为了通过调整底表层网的改变来提供面孔的新观点。我们测试了按固定基准顺序根据3D图像显示的图像配置方法。 Warping将延缩率降低70 $(从49.4ms到14.9ms),相应将缩放率超过多个GPU,同时将图像质量降低1$,使之适合作为端端到端视图的一部分。MA/DVAL 3D 应用程序。在 MA-D-D-DVAL 上找到。