Lip sync has emerged as a promising technique for generating mouth movements from audio signals. However, synthesizing a high-resolution and photorealistic virtual news anchor is still challenging. Lack of natural appearance, visual consistency, and processing efficiency are the main problems with existing methods. This paper presents a novel lip-sync framework specially designed for producing high-fidelity virtual news anchors. A pair of Temporal Convolutional Networks are used to learn the cross-modal sequential mapping from audio signals to mouth movements, followed by a neural rendering network that translates the synthetic facial map into a high-resolution and photorealistic appearance. This fully trainable framework provides end-to-end processing that outperforms traditional graphics-based methods in many low-delay applications. Experiments also show the framework has advantages over modern neural-based methods in both visual appearance and efficiency.
翻译:但是,合成高分辨率和摄影现实的虚拟新闻主播仍然具有挑战性。缺乏自然外观、视觉一致性和处理效率是现有方法的主要问题。本文介绍了一个新颖的唇合成框架,专门用来制作高不忠虚拟新闻主播。两组时空革命网络用来学习从音频信号到口音运动的跨模式相继制图,然后是将合成面貌图转换成高分辨率和光真化外观的神经转换网络。这个完全可训练的框架提供了终端到终端的处理,在许多低度应用中,它比传统的基于图形的方法更完善。实验还显示,这个框架在视觉和效率方面都比现代基于神经的方法优越。