In this paper, we work to bring telepresence to every desktop. Unlike commercial systems, personal 3D video conferencing systems must render high-quality videos while remaining financially and computationally viable for the average consumer. To this end, we introduce a capturing and rendering system that only requires 4 consumer-grade RGBD cameras and synthesizes high-quality free-viewpoint videos of users as well as their environments. Experimental results show that our system renders high-quality free-viewpoint videos without using object templates or heavy pre-processing. While not real-time, our system is fast and does not require per-video optimizations. Moreover, our system is robust to complex hand gestures and clothing, and it can generalize to new users. This work provides a strong basis for further optimization, and it will help bring telepresence to every desk in the near future. The code and dataset will be made available on our website https://mcmvmc.github.io/PersonalTelepresence/.
翻译:在本文中,我们致力于使每个桌面都具有远程交互功能。与商用系统不同,个人3D视频会议系统必须在保持财务和计算可行性的同时呈现高质量的视频,以适合普通消费者的使用。为此,我们介绍了一种捕捉和渲染系统,只需要4个普通的RGBD相机,就能合成高质量的用户自由视点视频以及他们的环境。实验结果表明,我们的系统可以在不使用对象模板或重型预处理的情况下呈现高质量的自由视点视频。虽然不是实时的,但我们的系统速度很快,不需要每个视频的优化。此外,我们的系统对于复杂的手势和服装很有 Robust,并且可以推广到新用户。本文为进一步优化提供了强有力的基础,并将有助于在不久的将来使每个桌面都具有远程交互功能。代码和数据集将在我们的网站https://mcmvmc.github.io/PersonalTelepresence/上提供。