This paper explores the bottleneck of feature representations of deep neural networks (DNNs), from the perspective of the complexity of interactions between input variables encoded in DNNs. To this end, we focus on the multi-order interaction between input variables, where the order represents the complexity of interactions. We discover that a DNN is more likely to encode both too simple interactions and too complex interactions, but usually fails to learn interactions of intermediate complexity. Such a phenomenon is widely shared by different DNNs for different tasks. This phenomenon indicates a cognition gap between DNNs and human beings, and we call it a representation bottleneck. We theoretically prove the underlying reason for the representation bottleneck. Furthermore, we propose a loss to encourage/penalize the learning of interactions of specific complexities, and analyze the representation capacities of interactions of different complexities.
翻译:本文从DNN编码输入变量之间复杂互动的角度,探讨深层神经网络特征表现的瓶颈。 为此,我们侧重于输入变量之间的多级互动,其顺序代表互动的复杂性。我们发现,DNN更有可能将过于简单的互动和过于复杂的互动结合起来,但通常无法了解中间复杂性的相互作用。不同DNN为不同任务广泛采用这种现象。这种现象表明DNN与人类之间的认知差距,我们称之为“代表瓶颈 ” 。我们理论上证明代表瓶颈的根本原因。此外,我们提议损失鼓励/整合对特定复杂互动的学习,并分析不同复杂互动的体现能力。