This paper introduces Associative Compression Networks (ACNs), a new framework for variational autoencoding with neural networks. The system differs from existing variational autoencoders in that the prior distribution used to model each code is conditioned on a similar code from the dataset. In compression terms this equates to sequentially transmitting the data using an ordering determined by proximity in latent space. As the prior need only account for local, rather than global variations in the latent space, the coding cost is greatly reduced, leading to rich, informative codes, even when autoregressive decoders are used. Experimental results on MNIST, CIFAR-10, ImageNet and CelebA show that ACNs can yield improved dataset compression relative to order-agnostic generative models, with an upper bound of 73.9 nats per image on binarized MNIST. They also demonstrate that ACNs learn high-level features such as object class, writing style, pose and facial expression, which can be used to cluster and classify the data, as well as to generate diverse and convincing samples.
翻译:本文介绍Associal自动编码网络(ACNs),这是一个用于神经网络的变异自动编码的新框架,该系统与现有的变异自动编码不同,因为用于模拟每个代码的先前分布以数据集的类似代码为条件。在压缩术语中,这相当于使用由隐蔽空间接近量决定的定序按顺序传送数据。由于先前的需要只考虑到潜伏空间的局部变化,而不是全球变化,因此编码成本大大降低,导致内容丰富、信息化的代码,即使在使用自动递增脱分器时也是如此。MNIST、CIFAR-10、图像Net和CelibA的实验结果显示,ACNs能够产生相对于定序-无序基因化模型的改进的数据集压缩,其上层为73.9纳特/每张力成像在MNISTS上。它们还表明,ACNs学习的高级特征,如对象类别、写作风格、面部和面部表达,可用于对数据进行分组和分类,以及生成多样化和令人信服的样本。