In recent years, Deep Learning methods have achieved state of the art performance in a vast range of machine learning tasks, including image classification and multilingual automatic text translation. These architectures are trained to solve machine learning tasks in an end-to-end fashion. In order to reach top-tier performance, these architectures often require a very large number of trainable parameters. There are multiple undesirable consequences, and in order to tackle these issues, it is desired to be able to open the black boxes of deep learning architectures. Problematically, doing so is difficult due to the high dimensionality of representations and the stochasticity of the training process. In this thesis, we investigate these architectures by introducing a graph formalism based on the recent advances in Graph Signal Processing (GSP). Namely, we use graphs to represent the latent spaces of deep neural networks. We showcase that this graph formalism allows us to answer various questions including: ensuring generalization abilities, reducing the amount of arbitrary choices in the design of the learning process, improving robustness to small perturbations added to the inputs, and reducing computational complexity
翻译:近些年来,深深学习方法在包括图像分类和多语言自动文本翻译在内的大量机器学习任务中达到了最新水平。这些架构经过培训,能够以端到端的方式解决机器学习任务。为了达到顶端的绩效,这些架构往往需要大量可培训的参数。有多种不良后果,为了解决这些问题,人们希望能够打开深层学习结构的黑盒子。 问题在于,由于演示的高度多面性和培训过程的随机性,很难做到这一点。 在论文中,我们根据图表信号处理(GSP)的最新进展,采用图表形式主义来调查这些架构。 也就是说,我们使用图表来代表深层神经网络的潜在空间。 我们展示了这个图表形式主义让我们能够回答各种问题,包括:确保普遍化能力,减少在设计学习过程中任意选择的数量,提高小扰动的强度,增加投入,降低计算的复杂性。