In animals, category learning enhances discrimination between stimuli close to the category boundary. This phenomenon, called categorical perception, was also empirically observed in artificial neural networks trained on classification tasks. In previous modeling works based on neuroscience data, we show that this expansion/compression is a necessary outcome of efficient learning. Here we extend our theoretical framework to artificial networks. We show that minimizing the Bayes cost (mean of the cross-entropy loss) implies maximizing the mutual information between the set of categories and the neural activities prior to the decision layer. Considering structured data with an underlying feature space of small dimension, we show that maximizing the mutual information implies (i) finding an appropriate projection space, and, (ii) building a neural representation with the appropriate metric. The latter is based on a Fisher information matrix measuring the sensitivity of the neural activity to changes in the projection space. Optimal learning makes this neural Fisher information follow a category-specific Fisher information, measuring the sensitivity of the category membership. Category learning thus induces an expansion of neural space near decision boundaries. We characterize the properties of the categorical Fisher information, showing that its eigenvectors give the most discriminant directions at each point of the projection space. We find that, unexpectedly, its maxima are in general not exactly at, but near, the class boundaries. Considering toy models and the MNIST dataset, we numerically illustrate how after learning the two Fisher information matrices match, and essentially align with the category boundaries. Finally, we relate our approach to the Information Bottleneck one, and we exhibit a bias-variance decomposition of the Bayes cost, of interest on its own.
翻译:在动物中,类别学习能够增强对接近类别边界刺激的区分能力。这一被称为"范畴感知"的现象,在面向分类任务训练的人工神经网络中也得到了实证观察。在我们先前基于神经科学数据的建模工作中,我们证明了这种扩张/压缩是高效学习的必然结果。本文将我们的理论框架扩展至人工网络。我们证明,最小化贝叶斯代价(交叉熵损失的均值)意味着最大化类别集合与决策层前神经活动之间的互信息。考虑到数据具有低维底层特征空间的结构,我们证明最大化互信息意味着:(i) 寻找合适的投影空间,以及(ii) 构建具有合适度量的神经表征。后者基于一个费舍尔信息矩阵,该矩阵度量神经活动对投影空间变化的敏感性。最优学习使得该神经费舍尔信息遵循一个类别特定的费舍尔信息,后者度量类别归属的敏感性。因此,类别学习会诱导决策边界附近神经空间的扩张。我们刻画了类别费舍尔信息的性质,证明其特征向量给出了投影空间各点处最具判别性的方向。我们发现,出乎意料的是,其最大值通常并不精确地位于类别边界上,而是靠近边界。通过玩具模型和MNIST数据集,我们数值化地展示了学习后两个费舍尔信息矩阵如何匹配,并基本与类别边界对齐。最后,我们将本方法与信息瓶颈方法联系起来,并展示了贝叶斯代价的偏差-方差分解,这一分解本身具有独立意义。