Deep learning of artificial neural networks (ANNs) is creating highly functional processes that are, unfortunately, nearly as hard to interpret as their biological counterparts. Identification of functional modules in natural brains plays an important role in cognitive and neuroscience alike, and can be carried out using a wide range of technologies such as fMRI, EEG/ERP, MEG, or calcium imaging. However, we do not have such robust methods at our disposal when it comes to understanding functional modules in artificial neural networks. Ideally, understanding which parts of an artificial neural network perform what function might help us to address a number of vexing problems in ANN research, such as catastrophic forgetting and overfitting. Furthermore, revealing a network's modularity could improve our trust in them by making these black boxes more transparent. Here, we introduce a new information-theoretic concept that proves useful in understanding and analyzing a network's functional modularity: the relay information $I_R$. The relay information measures how much information groups of neurons that participate in a particular function (modules) relay from inputs to outputs. Combined with a greedy search algorithm, relay information can be used to identify computational modules in neural networks. We also show that the functionality of modules correlates with the amount of relay information they carry.
翻译:对人工神经网络(ANNs)的深入学习正在创造高度功能化的过程,不幸的是,这些过程几乎很难与生物对等体一样解释。自然大脑中功能模块的识别在认知和神经科学中都起着重要作用,并且可以使用广泛的技术,例如FMRI、EEG/ERP、MEG或钙成像等技术来进行。然而,当了解人工神经网络中的功能模块时,我们没有这种强有力的方法可供我们使用。理想的情况是,了解人工神经网络中哪些部分的功能可以帮助我们解决ANN研究中一些棘手的问题,例如灾难性的遗忘和过度装配。此外,披露网络的模块性可以通过使这些黑盒更加透明来增进我们对这些模块的信任。在这里,我们引入一种新的信息理论概念,在理解和分析网络功能模块时证明有用:中继信息 $I_R$。 中继信息测量参与特定功能(模块)从输入到输出中的信息组有多少可以帮助我们解决一些棘手的问题。结合一个贪婪的搜索算法,中继信息模块也用于计算。</s>