Deep neural networks unlocked a vast range of new applications by solving tasks of which many were previously deemed as reserved to higher human intelligence. One of the developments enabling this success was a boost in computing power provided by special purpose hardware, such as graphic or tensor processing units. However, these do not leverage fundamental features of neural networks like parallelism and analog state variables. Instead, they emulate neural networks relying on binary computing, which results in unsustainable energy consumption and comparatively low speed. Fully parallel and analogue hardware promises to overcome these challenges, yet the impact of analogue neuron noise and its propagation, i.e. accumulation, threatens rendering such approaches inept. Here, we determine for the first time the propagation of noise in deep neural networks comprising noisy nonlinear neurons in trained fully connected layers. We study additive and multiplicative as well as correlated and uncorrelated noise, and develop analytical methods that predict the noise level in any layer of symmetric deep neural networks or deep neural networks trained with back propagation. We find that noise accumulation is generally bound, and adding additional network layers does not worsen the signal to noise ratio beyond a limit. Most importantly, noise accumulation can be suppressed entirely when neuron activation functions have a slope smaller than unity. We therefore developed the framework for noise in fully connected deep neural networks implemented in analog systems, and identify criteria allowing engineers to design noise-resilient novel neural network hardware.
翻译:深心神经网络通过解决许多以前被认为保留给更高人类智慧的新任务而打开了广泛的新应用。 促成这一成功的发展之一是特殊用途硬件,如图形或高压处理器提供的计算机动力的增强。 但是,这些并不利用神经网络的基本特征,如平行和模拟状态变量。 相反,它们效仿依赖双轨计算而导致不可持续能源消耗和相对较低速度的神经网络。 完全平行和模拟的硬件有望克服这些挑战,而模拟神经神经噪音及其传播,即累积的影响则有可能使这些方法难以采用。 在这里,我们首次确定由经过训练的完全相连的层中噪音的非线性神经元组成的深线性网络的噪音传播。 我们研究添加和多复制性以及相关性和非线性噪音,并开发出分析方法来预测任何对称深度的深层神经网络的噪音水平,或经过后期传播训练的深层神经网络。我们发现,噪音的累积及其传播,即累积,威胁着这种方法的传播,从而有可能使这种方法变得不易。在这里,我们第一次确定由经过训练的非线性神经网络传播的神经网络传播的噪音网络传播,因此使得信号变得无法在深度结构结构中变小。 。最重要的是,因此,噪音网络可以完全被固定地固定。最重要的是进行。 最重要的是,因此,噪音累积的网络被固定。