We present LR-GAN: an adversarial image generation model which takes scene structure and context into account. Unlike previous generative adversarial networks (GANs), the proposed GAN learns to generate image background and foregrounds separately and recursively, and stitch the foregrounds on the background in a contextually relevant manner to produce a complete natural image. For each foreground, the model learns to generate its appearance, shape and pose. The whole model is unsupervised, and is trained in an end-to-end manner with gradient descent methods. The experiments demonstrate that LR-GAN can generate more natural images with objects that are more human recognizable than DCGAN.
翻译:我们展示了LR-GAN:一种将场景结构和背景考虑在内的对抗性图像生成模型。与以往的基因对抗网络(GANs)不同,拟议的GAN学会了分别生成图像背景和前景,并会回溯并重现,以与背景相关的方式在背景上对前景进行缝合,以生成完整的自然图像。对于每个前景,模型学会生成其外观、形状和外观。整个模型不受监督,并且以梯度下移方法以端至端的方式接受培训。实验表明,LR-GAN能够生成比DCGAN更人类可识别的物体的更多自然图像。