Deep generative approaches have recently made considerable progress in image inpainting by introducing structure priors. Due to the lack of proper interaction with image texture during structure reconstruction, however, current solutions are incompetent in handling the cases with large corruptions, and they generally suffer from distorted results. In this paper, we propose a novel two-stream network for image inpainting, which models the structure-constrained texture synthesis and texture-guided structure reconstruction in a coupled manner so that they better leverage each other for more plausible generation. Furthermore, to enhance the global consistency, a Bi-directional Gated Feature Fusion (Bi-GFF) module is designed to exchange and combine the structure and texture information and a Contextual Feature Aggregation (CFA) module is developed to refine the generated contents by region affinity learning and multi-scale feature aggregation. Qualitative and quantitative experiments on the CelebA, Paris StreetView and Places2 datasets demonstrate the superiority of the proposed method. Our code is available at https://github.com/Xiefan-Guo/CTSDG.
翻译:最近,由于在结构重建期间没有与图像纹理进行适当的互动,目前的解决办法无法处理大量腐败的案件,而且一般都受到扭曲的结果的影响。在本文件中,我们提议建立一个新型的两流图像涂料网络,以结构限制的纹理合成和纹理指导结构的重建为模型,同时以各种方式对结构限制的纹理合成和纹理指导结构的重建进行模型,使它们更好地为更合理的一代相互利用。此外,为了提高全球一致性,设计了一个双向Getature Fusion(Bi-GFF)模块,以交换和合并结构和纹理信息,并开发一个背景特征聚合模块,以完善区域亲和多尺度特征聚合所生成的内容。CelebA、巴黎街道View和Places2数据集的定性和定量实验展示了拟议方法的优越性。我们的代码可在https://github.com/Xifan-Guo/CTSDG上查阅。