Face-swapping models have been drawing attention for their compelling generation quality, but their complex architectures and loss functions often require careful tuning for successful training. We propose a new face-swapping model called `Smooth-Swap', which excludes complex handcrafted designs and allows fast and stable training. The main idea of Smooth-Swap is to build smooth identity embedding that can provide stable gradients for identity change. Unlike the one used in previous models trained for a purely discriminative task, the proposed embedding is trained with a supervised contrastive loss promoting a smoother space. With improved smoothness, Smooth-Swap suffices to be composed of a generic U-Net-based generator and three basic loss functions, a far simpler design compared with the previous models. Extensive experiments on face-swapping benchmarks (FFHQ, FaceForensics++) and face images in the wild show that our model is also quantitatively and qualitatively comparable or even superior to the existing methods.
翻译:面部擦洗模型一直吸引人们关注其惊人的一代质量,但是其复杂的结构和损失功能往往需要仔细调整才能成功培训。我们提出了一个新的面部擦洗模型,名为“Smooth-Swap ”,它排除了复杂的手工设计,可以进行快速稳定的培训。光滑-Swap 的主要想法是建立光滑的身份嵌入,为身份变化提供稳定的梯度。与以前为纯粹歧视性任务而培训的模型不同,提议的嵌入过程经过监督的对比损失培训,促进更平滑的空间。随着光滑,光滑Swap足以由通用的U-Net型发电机和三个基本损失功能组成,与以前的模型相比设计要简单得多。关于面部擦拭基准(FFHQ,FaceForensics++)和野生图像的广泛实验显示,我们的模型在数量和质量上也具有可比性,甚至优于现有方法。