Augmented reality applications have rapidly spread across online platforms, allowing consumers to virtually try-on a variety of products, such as makeup, hair dying, or shoes. However, parametrizing a renderer to synthesize realistic images of a given product remains a challenging task that requires expert knowledge. While recent work has introduced neural rendering methods for virtual try-on from example images, current approaches are based on large generative models that cannot be used in real-time on mobile devices. This calls for a hybrid method that combines the advantages of computer graphics and neural rendering approaches. In this paper we propose a novel framework based on deep learning to build a real-time inverse graphics encoder that learns to map a single example image into the parameter space of a given augmented reality rendering engine. Our method leverages self-supervised learning and does not require labeled training data which makes it extendable to many virtual try-on applications. Furthermore, most augmented reality renderers are not differentiable in practice due to algorithmic choices or implementation constraints to reach real-time on portable devices. To relax the need for a graphics-based differentiable renderer in inverse graphics problems, we introduce a trainable imitator module. Our imitator is a generative network that learns to accurately reproduce the behavior of a given non-differentiable renderer. We propose a novel rendering sensitivity loss to train the imitator, which ensures that the network learns an accurate and continuous representation for each rendering parameter. Our framework enables novel applications where consumers can virtually try-on a novel unknown product from an inspirational reference image on social media. It can also be used by graphics artists to automatically create realistic rendering from a reference product image.
翻译:强化的现实应用在网上平台上迅速扩散,使消费者能够几乎尝试各种产品,如化妆、发毛死亡或鞋等。然而,合成某一产品现实图像的合成器仍然是一项艰巨的任务,需要专家知识。虽然最近的工作引入了神经转换方法,从示例图像进行虚拟试验,但当前的方法基于大型基因化模型,无法在移动设备上实时使用。这要求一种混合方法,将计算机现实化图形和神经生化方法的优势结合起来。在这份文件中,我们提议了一个基于深层学习的新框架,以建立实时的反向图形应用,以建立实时的虚拟图形化应用;虚拟化虚拟化的图形编码,学会将一个单一的示例图像图像映射到一个特定增强的现实化引擎的参数空间。我们的方法利用了自我超强的学习方法,而不需要贴标签的培训数据,使得它可以推广到许多虚拟试镜应用程序。此外,由于对虚拟化和内装工具实施的限制,在实践上是没有区别的。为了放松对基于图形的引用的精确度引用的图像和直观参考框架的需求,我们也可以将一个不易读的图像转换成一个不易变的图像的网络,我们在图形中将一个图像模型中将一个图像变的模型中将一个图像变的模型转换成一个不动的模型的模型。