In this work, we investigate the use of three information-theoretic quantities -- entropy, mutual information with the class variable, and a class selectivity measure based on Kullback-Leibler divergence -- to understand and study the behavior of already trained fully-connected feed-forward neural networks. We analyze the connection between these information-theoretic quantities and classification performance on the test set by cumulatively ablating neurons in networks trained on MNIST, FashionMNIST, and CIFAR-10. Our results parallel those recently published by Morcos et al., indicating that class selectivity is not a good indicator for classification performance. However, looking at individual layers separately, both mutual information and class selectivity are positively correlated with classification performance, at least for networks with ReLU activation functions. We provide explanations for this phenomenon and conclude that it is ill-advised to compare the proposed information-theoretic quantities across layers. Furthermore, we show that cumulative ablation of neurons with ascending or descending information-theoretic quantities can be used to formulate hypotheses regarding the joint behavior of multiple neurons, such as redundancy and synergy, with comparably low computational cost. We also draw connections to the information bottleneck theory for neural networks.
翻译:在这项工作中,我们调查使用三种信息理论数量 -- -- 昆虫、与阶级变量的相互信息,以及基于库尔贝克-利伯尔差异的阶级选择性措施 -- -- 来理解和研究已经受过训练的完全连接的向神经网络的进化传感网络的行为。我们分析了在MNIST、时装MIS和CIFAR-10培训的网络中累积消化神经元所设定的测试信息理论数量和分类性能的分类性能之间的关联。我们的结果与Morcos等人最近公布的结果相平行,表明阶级选择性不是分类性能的好指标。然而,分别看各个层次,相互信息和阶级选择性与分类性能有正相关关系,至少对于具有RELU激活功能的网络而言。我们对这种现象作出解释,并得出结论认为,对跨层次的拟议信息理论数量进行对比是不明智的。此外,我们表明,神经元与信息升降或降序数量之间的累积性关系可以用来为多个神经神经联合行为的假设,例如与神经元的理论连接,我们也可以进行低量的计算。