Image-based virtual try-on techniques have shown great promise for enhancing the user-experience and improving customer satisfaction on fashion-oriented e-commerce platforms. However, existing techniques are currently still limited in the quality of the try-on results they are able to produce from input images of diverse characteristics. In this work, we propose a Context-Driven Virtual Try-On Network (C-VTON) that addresses these limitations and convincingly transfers selected clothing items to the target subjects even under challenging pose configurations and in the presence of self-occlusions. At the core of the C-VTON pipeline are: (i) a geometric matching procedure that efficiently aligns the target clothing with the pose of the person in the input images, and (ii) a powerful image generator that utilizes various types of contextual information when synthesizing the final try-on result. C-VTON is evaluated in rigorous experiments on the VITON and MPV datasets and in comparison to state-of-the-art techniques from the literature. Experimental results show that the proposed approach is able to produce photo-realistic and visually convincing results and significantly improves on the existing state-of-the-art.
翻译:以图像为基础的虚拟试镜技术在加强用户经验和提高客户对时装电子商务平台满意度方面显示出巨大的希望,但是,现有技术目前仍然在它们能够从不同特点的输入图像中产生的试镜结果质量方面受到限制;在这项工作中,我们提议建立 " 环境驱动虚拟试镜网络 " (C-VTON),以克服这些局限性,并令人信服地将选定的服装项目转让给目标对象,即使是在具有挑战性配置和自我封闭的情况下也是如此。 C-VTON管道的核心是:(一) 几何匹配程序,有效地将目标服装与输入图像中的人的面容相匹配,以及(二) 强大的图像生成器,在综合最后试镜结果时利用各种背景信息。C-VTON在对VITON和MPV数据集进行严格试验,并与文献中的最新技术进行比较,对C-VTON进行了评价。实验结果表明,拟议的方法能够产生摄影现实和视觉的令人信服的结果,并大大改进了现有状态。