This paper aims to theoretically analyze the complexity of feature transformations encoded in DNNs with ReLU layers. We propose metrics to measure three types of complexities of transformations based on the information theory. We further discover and prove the strong correlation between the complexity and the disentanglement of transformations. Based on the proposed metrics, we analyze two typical phenomena of the change of the transformation complexity during the training process, and explore the ceiling of a DNN's complexity. The proposed metrics can also be used as a loss to learn a DNN with the minimum complexity, which also controls the over-fitting level of the DNN and influences adversarial robustness, adversarial transferability, and knowledge consistency. Comprehensive comparative studies have provided new perspectives to understand the DNN.
翻译:本文旨在从理论上分析DNN和RELU层次编码的特征转换的复杂性,我们根据信息理论提出衡量三种类型变异复杂性的指标,我们进一步发现和证明变异的复杂性和分解性之间的密切关系。根据拟议的衡量标准,我们分析了培训过程中变异复杂性变化的两个典型现象,并探讨了DNN的复杂程度的上限。拟议的衡量标准也可以作为一种损失,用来学习具有最低复杂性的DNN,这也控制了DN的过度配置水平,并影响了对抗性强健性、对抗性转移性和知识一致性。全面的比较研究为理解DNN提供了新的视角。