In this work we seek to bridge the concepts of topographic organization and equivariance in neural networks. To accomplish this, we introduce the Topographic VAE: a novel method for efficiently training deep generative models with topographically organized latent variables. We show that such a model indeed learns to organize its activations according to salient characteristics such as digit class, width, and style on MNIST. Furthermore, through topographic organization over time (i.e. temporal coherence), we demonstrate how predefined latent space transformation operators can be encouraged for observed transformed input sequences -- a primitive form of unsupervised learned equivariance. We demonstrate that this model successfully learns sets of approximately equivariant features (i.e. "capsules") directly from sequences and achieves higher likelihood on correspondingly transforming test sequences. Equivariance is verified quantitatively by measuring the approximate commutativity of the inference network and the sequence transformations. Finally, we demonstrate approximate equivariance to complex transformations, expanding upon the capabilities of existing group equivariant neural networks.
翻译:在这项工作中,我们试图将地形组织的概念和神经网络的均匀性概念连接起来。为了实现这一点,我们引入了地形VAE:一种高效培训具有地形组织潜伏变量的深基因模型的新颖方法。我们表明,这种模型确实学会根据数字级、宽度和MNIST的风格等显著特征来组织其启动活动。此外,通过随着时间的推移的地形组织(即时间一致性),我们展示了如何鼓励预定义的潜在空间转换操作器对观察到的转换输入序列 -- -- 一种原始的、未经监督的、有知识的等同性形式。我们证明,这一模型直接从序列中学习了几套近等异性特征(即“囊”)并取得了相应的转换测试序列的更大可能性。量化差异是通过测量推断网络和序列转换的近似通性来加以验证的。最后,我们显示了对复杂变异性的近等性,以现有群体等性网络的能力为基础扩大。