We show that pre-trained Generative Adversarial Networks (GANs) such as StyleGAN and BigGAN can be used as a latent bank to improve the performance of image super-resolution. While most existing perceptual-oriented approaches attempt to generate realistic outputs through learning with adversarial loss, our method, Generative LatEnt bANk (GLEAN), goes beyond existing practices by directly leveraging rich and diverse priors encapsulated in a pre-trained GAN. But unlike prevalent GAN inversion methods that require expensive image-specific optimization at runtime, our approach only needs a single forward pass for restoration. GLEAN can be easily incorporated in a simple encoder-bank-decoder architecture with multi-resolution skip connections. Employing priors from different generative models allows GLEAN to be applied to diverse categories (\eg~human faces, cats, buildings, and cars). We further present a lightweight version of GLEAN, named LightGLEAN, which retains only the critical components in GLEAN. Notably, LightGLEAN consists of only 21% of parameters and 35% of FLOPs while achieving comparable image quality. We extend our method to different tasks including image colorization and blind image restoration, and extensive experiments show that our proposed models perform favorably in comparison to existing methods. Codes and models are available at https://github.com/open-mmlab/mmediting.
翻译:我们展示了经过培训的GAN(StyleGAN)和BigGAN(BigGAN)等经过培训的基因反转网络(GANs)可以用作提高图像超分辨率性能的潜在库库。虽然大多数现有的观念导向方法都试图通过学习对抗性损失而产生现实产出,但我们的方法,即General LatEnt bANk(GLEAN),直接利用在经过培训的GAN(GAN)中包含的丰富和多样的前科。但是,与普通GAN Inversion(SylegGAN)的通用方法不同,在运行时需要花费昂贵的图像特优异性优化,我们的方法只需要一个单一的远端传球。GLANEAN(GLAN)可以很容易地融入一个简单的编码-银行分解器结构,而多分辨率的连接会跳过。从不同的基因模型使用以前的GLANEAN(GLAN),可以将GLAN(GLAN)应用到不同的类别(GLAN)应用。我们现有的图像分析模型可以扩展的模型。我们现有的图像分析方法,我们可以用了。