The information bottleneck framework provides a systematic approach to learning representations that compress nuisance information in the input and extract semantically meaningful information about predictions. However, the choice of a prior distribution that fixes the dimensionality across all the data can restrict the flexibility of this approach for learning robust representations. We present a novel sparsity-inducing spike-slab categorical prior that uses sparsity as a mechanism to provide the flexibility that allows each data point to learn its own dimension distribution. In addition, it provides a mechanism for learning a joint distribution of the latent variable and the sparsity and hence can account for the complete uncertainty in the latent space. Through a series of experiments using in-distribution and out-of-distribution learning scenarios on the MNIST, CIFAR-10, and ImageNet data, we show that the proposed approach improves accuracy and robustness compared to traditional fixed-dimensional priors, as well as other sparsity induction mechanisms for latent variable models proposed in the literature.
翻译:信息瓶颈框架提供了一种系统性的学习表现方法,在输入中压缩干扰信息,并提取关于预测的精度有意义的信息。然而,选择一种先前的分布方法,以修正所有数据的维度,可能会限制这一方法在学习强健的表述方面的灵活性。我们介绍了一种新颖的聚度诱导螺旋丝状板,在之前将聚度作为一种提供灵活性的机制,使每个数据点能够了解其自身的维度分布。此外,它提供了一个机制,用于学习潜在变量和孔隙的共同分布,从而可以说明潜在空间的完全不确定性。通过在MNIST、CIFAR-10和图像网络数据上进行一系列分配和分配外学习的实验,我们表明,拟议的方法与传统的固定维的前期数据相比,提高了准确性和稳健性,以及文献中提议的关于潜在变量模型的其他快速感应感机制。