Intuitively editing the appearance of materials from a single image is a challenging task given the complexity of the interactions between light and matter, and the ambivalence of human perception. This problem has been traditionally addressed by estimating additional factors of the scene like geometry or illumination, thus solving an inverse rendering problem and subduing the final quality of the results to the quality of these estimations. We present a single-image appearance editing framework that allows us to intuitively modify the material appearance of an object by increasing or decreasing high-level perceptual attributes describing such appearance (e.g., glossy or metallic). Our framework takes as input an in-the-wild image of a single object, where geometry, material, and illumination are not controlled, and inverse rendering is not required. We rely on generative models and devise a novel architecture with Selective Transfer Unit (STU) cells that allow to preserve the high-frequency details from the input image in the edited one. To train our framework we leverage a dataset with pairs of synthetic images rendered with physically-based algorithms, and the corresponding crowd-sourced ratings of high-level perceptual attributes. We show that our material editing framework outperforms the state of the art, and showcase its applicability on synthetic images, in-the-wild real-world photographs, and video sequences.
翻译:从单一图像中直接编辑材料外观是一项艰巨的任务,因为光和物质之间的相互作用十分复杂,而且人类感知的矛盾性也十分复杂。 这个问题传统上是通过估算几何或光照等场景的更多因素来解决的,从而解决了反向转化问题,将结果的最终质量降低到这些估计的质量。 我们提出了一个单一的外观编辑框架,使我们能够通过增加或减少描述这种外观的高层次感官特征(如光和物质或金属)来直观地修改一个对象的材料外观。 我们的框架是作为输入一个单一对象(即不控制几何、材料和光照的)的全方位图像的全方位图像,而不需要反面图像。 我们依靠的是变色模型,并与选择性传输股(STU)的细胞设计了一个新的结构,从而能够保存从编辑的输入图像中获取的高频细节。 为了培训我们的框架,我们利用一组合成图像(如物理算法或金属)作为输入的组合图像。 我们的框架将一个单一对象的全方位图像作为输入输入,其中的图像、材料、材料、材料和光质图像的可应用性图像,我们在高端的图像的图像上显示高端的图像的图像。