Generating new images with desired properties (e.g. new view/poses) from source images has been enthusiastically pursued recently, due to its wide range of potential applications. One way to ensure high-quality generation is to use multiple sources with complementary information such as different views of the same object. However, as source images are often misaligned due to the large disparities among the camera settings, strong assumptions have been made in the past with respect to the camera(s) or/and the object in interest, limiting the application of such techniques. Therefore, we propose a new general approach which models multiple types of variations among sources, such as view angles, poses, facial expressions, in a unified framework, so that it can be employed on datasets of vastly different nature. We verify our approach on a variety of data including humans bodies, faces, city scenes and 3D objects. Both the qualitative and quantitative results demonstrate the better performance of our method than the state of the art.
翻译:最近,由于源图像具有各种潜在应用,因此热切地从源图像中生成有预期特性的新图像(例如,新的视图/位置),最近由于源图像具有广泛的潜在应用,因此得到了热烈的探索。确保高质量生成的一种方法是使用多种来源,同时提供补充信息,例如同一对象的不同观点。然而,由于相机设置之间差异很大,源图像往往不吻合,因此过去对相机或(和)感兴趣的对象作了强烈的假设,限制了这些技术的应用。因此,我们提议一种新的通用方法,在统一的框架内模拟各种来源之间的多种差异,例如视图角度、外观、面部表情表达,以便能够在性质截然不同的数据集中使用。我们核查了我们关于各种数据的方法,包括人体、面部、城市场景和3D天体。质量和数量结果都表明我们方法的绩效优于艺术状态。