Despite significant advances in the field of deep learning in applications to various fields, explaining the inner processes of deep learning models remains an important and open question. The purpose of this article is to describe and substantiate the geometric and topological view of the learning process of neural networks. Our attention is focused on the internal representation of neural networks and on the dynamics of changes in the topology and geometry of the data manifold on different layers. We also propose a method for assessing the generalizing ability of neural networks based on topological descriptors. In this paper, we use the concepts of topological data analysis and intrinsic dimension, and we present a wide range of experiments on different datasets and different configurations of convolutional neural network architectures. In addition, we consider the issue of the geometry of adversarial attacks in the classification task and spoofing attacks on face recognition systems. Our work is a contribution to the development of an important area of explainable and interpretable AI through the example of computer vision.
翻译:尽管在各个领域应用的深层学习领域取得了重大进展,但解释深层学习模型的内部过程仍是一个重要和未决问题。本篇文章的目的是描述和证实神经网络学习过程的几何和地形学观点。我们关注的重点是神经网络的内部代表性和不同层次数据表层和几何变化动态。我们还提议了一种方法,用以评估基于地形描述器的神经网络的普遍化能力。在本文件中,我们使用了地形数据分析和内在层面的概念,并提出了关于不同数据集和同源神经网络结构不同配置的广泛实验。此外,我们还审议了分类任务中的对抗性攻击的几何学问题和对面部识别系统进行模拟攻击的问题。我们的工作有助于以计算机视觉为例,发展一个可解释和可解释的AI的重要领域。