Full characterization of the spectral behavior of generative models based on neural networks remains an open issue. Recent research has focused heavily on generative adversarial networks and the high-frequency discrepancies between real and generated images. The current solution to avoid this is to either replace transposed convolutions with bilinear up-sampling or add a spectral regularization term in the generator. It is well known that Variational Autoencoders (VAEs) also suffer from these issues. In this work, we propose a simple 2D Fourier transform-based spectral regularization loss for the VAE and show that it can achieve results equal to, or better than, the current state-of-the-art in frequency-aware losses for generative models. In addition, we experiment with altering the up-sampling procedure in the generator network and investigate how it influences the spectral performance of the model. We include experiments on synthetic and real data sets to demonstrate our results.
翻译:以神经网络为基础的基因模型的光谱行为的全面定性仍是一个尚未解决的问题。 最近的研究主要侧重于基因对抗网络以及真实图像和生成图像之间的高频差。 目前避免这种情况的解决方案是要么用双线上取样取代转换的演进,要么在发电机中添加一个光谱正规化术语。众所周知,变化式自动调控器(VAEs)也受到这些问题的影响。在这项工作中,我们提议为VAE提供一个简单的 2D Fourier 变异式光谱规范损失,并表明它能够取得相当于或更好的效果,相当于或比当前对基因模型最先进的频率觉测损失。此外,我们尝试改变发电机网络的升级程序,并研究它如何影响模型的光谱性能。我们把对合成和真实数据集的实验包括在内,以展示我们的结果。