Artificial and biological agents cannon learn given completely random and unstructured data. The structure of data is encoded in the metric relationships between data points. In the context of neural networks, neuronal activity within a layer forms a representation reflecting the transformation that the layer implements on its inputs. In order to utilize the structure in the data in a truthful manner, such representations should reflect the input distances and thus be continuous and isometric. Supporting this statement, recent findings in neuroscience propose that generalization and robustness are tied to neural representations being continuously differentiable. In machine learning, most algorithms lack robustness and are generally thought to rely on aspects of the data that differ from those that humans use, as is commonly seen in adversarial attacks. During cross-entropy classification, the metric and structural properties of network representations are usually broken both between and within classes. This side effect from training can lead to instabilities under perturbations near locations where such structure is not preserved. One of the standard solutions to obtain robustness is to add ad hoc regularization terms, but to our knowledge, forcing representations to preserve the metric structure of the input data as a stabilising mechanism has not yet been studied. In this work, we train neural networks to perform classification while simultaneously maintaining within-class metric structure, leading to isometric within-class representations. Such network representations turn out to be beneficial for accurate and robust inference. By stacking layers with this property we create a network architecture that facilitates hierarchical manipulation of internal neural representations. Finally, we verify that isometric regularization improves the robustness to adversarial attacks on MNIST.
翻译:人工和生物剂大炮的学习完全随机和不结构化的数据。数据结构在数据点之间的测量关系中进行编码。在神经网络中,层内的神经活动形成一个代表,反映该层投入的转变。为了真实地利用数据结构,这种表述应反映输入距离,因而是连续的和测量的。为了支持这一陈述,神经科学的最新发现表明,一般化和稳健性与神经表达相联,不断有差异。在机器学习中,大多数算法缺乏稳健性,一般认为依赖与人类使用的数据不同的数据方面。在对立攻击中常见的,层内的神经活动反映了该层的转变。在交叉性分类中,网络表述的衡量和结构通常在班级之间和班级内部破裂。这种培训的侧面效应可能导致在不保存结构的附近扰动下出现不稳性。 获得稳健性的一个标准解决方案是增加临时的正规化条件,但对于我们的知识而言,要求通过保持矩阵结构的准确性结构来维护与人类使用的不同方面的数据。在结构内部进行精确性分类中进行这种结构的调整,我们一直在研究,在进行这种结构内部结构内部结构中进行这种结构结构的调整,在进行这种结构内部结构内部结构结构内进行这种结构的调整。