Despite unconditional feature inversion being the foundation of many image synthesis applications, training an inverter demands a high computational budget, large decoding capacity and imposing conditions such as autoregressive priors. To address these limitations, we propose the use of adversarially robust representations as a perceptual primitive for feature inversion. We train an adversarially robust encoder to extract disentangled and perceptually-aligned image representations, making them easily invertible. By training a simple generator with the mirror architecture of the encoder, we achieve superior reconstruction quality and generalization over standard models. Based on this, we propose an adversarially robust autoencoder and demonstrate its improved performance on style transfer, image denoising and anomaly detection tasks. Compared to recent ImageNet feature inversion methods, our model attains improved performance with significantly less complexity.
翻译:尽管无条件的特征转换是许多图像合成应用程序的基础,培训一个倒置器需要很高的计算预算、大量的解码能力和强制条件,如自动递减前科等。为了解决这些限制,我们提议使用对抗性强的演示作为特征转换的原始概念。我们训练一个对抗性强的编码器,以提取分解和感知相近的图像表达,使其容易被翻转。通过用编码器的镜像结构培训一个简单的生成器,我们实现了较高的重建质量和比标准模型的概括化。基于这一点,我们提议了一个对抗性强的自动编码器,并展示了其在风格转换、图像除色和异常检测任务方面的改进性能。与最近的图像网络特征转换方法相比,我们的模型在改进性能的同时,复杂性也大大降低。