Classification is one of the major tasks that deep learning is successfully tackling. Categorization is also a fundamental cognitive ability. A well-known perceptual consequence of categorization in humans and other animals, called categorical perception, is characterized by a within-category compression and a between-category separation: two items, close in input space, are perceived closer if they belong to the same category than if they belong to different categories. Elaborating on experimental and theoretical results in cognitive science, here we study categorical effects in artificial neural networks. Our formal and numerical analysis provides insights into the geometry of the neural representation in deep layers, with expansion of space near category boundaries and contraction far from category boundaries. We investigate categorical representation by using two complementary approaches: one mimics experiments in psychophysics and cognitive neuroscience by means of morphed continua between stimuli of different categories, while the other introduces a categoricality index that quantifies the separability of the classes at the population level (a given layer in the neural network). We show on both shallow and deep neural networks that category learning automatically induces categorical perception. We further show that the deeper a layer, the stronger the categorical effects. An important outcome of our analysis is to provide a coherent and unifying view of the efficacy of different heuristic practices of the dropout regularization technique. Our views, which find echoes in the neuroscience literature, insist on the differential role of noise as a function of the level of representation and in the course of learning: noise injected in the hidden layers gets structured according to the organization of the categories, more variability being allowed within a category than across classes.
翻译:分类是深层次学习成功处理的主要任务之一。 分类也是基本的认知能力。 人类和其他动物分类的一个众所周知的概念性后果,称为绝对感知,其特点是类内压缩和类别间分离:两个项目,接近投入空间,如果属于同一类别,则被认为更接近于属于不同类别。 阐述认知科学的实验和理论结果,我们在这里研究人工神经网络中的直线效应。 我们的正式和数字分析为深层神经结构的几何分布提供了深刻的洞察力,其间空间靠近类别界限,缩小到远离类别界限的缩小。 我们通过两种互补方法来调查绝对代表性:一个在精神物理和认知神经科学方面的模拟实验,其方式是不同类别之间的变相调,而另一个则引入了明确性指数,该指数使人口层次(神经网络中的一种给定层)的分辨能力变得分化。 我们展示了浅层和深层神经网络的几何分解,该类别可以自动引起直截的分辨的分辨认识。 我们通过两种互补的分层的分层分析,我们更深刻的分层的分层的分层的分层的分层分析,让我们的分解到分层的分层的分层分析。 更深层次的分层的分层的分层的分层的分层的分层的分层的分层的分层的分解,是分层的分层的分层的分层的分层的分层的分层的分层的分层的分层, 。