Image-based virtual try-on provides the capacity to transfer a clothing item onto a photo of a given person, which is usually accomplished by warping the item to a given human pose and adjusting the warped item to the person. However, the results of real-world synthetic images (e.g., selfies) from the previous method is not realistic because of the limitations which result in the neck being misrepresented and significant changes to the style of the garment. To address these challenges, we propose a novel method to solve this unique issue, called VITON-CROP. VITON-CROP synthesizes images more robustly when integrated with random crop augmentation compared to the existing state-of-the-art virtual try-on models. In the experiments, we demonstrate that VITON-CROP is superior to VITON-HD both qualitatively and quantitatively.
翻译:以图像为基础的虚拟试镜提供了将衣物物品转移到某个人的照片上的能力,通常通过将衣物物品切换成给定人的姿势和将扭曲的物品调整成人来完成。然而,以前方法中真实世界合成图像(如自相)的结果并不现实,因为其局限性导致颈部被歪曲,服装风格也发生了重大变化。为了应对这些挑战,我们提出了一个解决这一独特问题的新方法,称为VITON-CROP。 VITON-CROP在与随机作物增殖相结合时,与现有最先进的虚拟试镜模型相比,将图像合成得更加稳健。 在实验中,我们证明VITON-CROP在质量和数量上都优于VITON-HD。