Deep neural networks unlocked a vast range of new applications by solving tasks of which many were previouslydeemed as reserved to higher human intelligence. One of the developments enabling this success was a boost incomputing power provided by special purpose hardware, such as graphic or tensor processing units. However,these do not leverage fundamental features of neural networks like parallelism and analog state variables.Instead, they emulate neural networks relying on computing power, which results in unsustainable energyconsumption and comparatively low speed. Fully parallel and analogue hardware promises to overcomethese challenges, yet the impact of analogue neuron noise and its propagation, i.e. accumulation, threatensrendering such approaches inept. Here, we analyse for the first time the propagation of noise in paralleldeep neural networks comprising noisy nonlinear neurons. We develop an analytical treatment for both,symmetric networks to highlight the underlying mechanisms, and networks trained with back propagation.We find that noise accumulation is generally bound, and adding additional network layers does not worsenthe signal to noise ratio beyond this limit. Most importantly, noise accumulation can be suppressed entirelywhen neuron activation functions have a slope smaller than unity. We therefore developed the frameworkfor noise of deep neural networks implemented in analog systems, and identify criteria allowing engineers todesign noise-resilient novel neural network hardware.
翻译:深心神经网络通过解决许多以前被认为是人类更高智慧所保留的任务,打开了广泛的新应用。 促成这一成功的一项发展是特殊用途硬件,如图形或高压处理器提供的加速计算能力。 然而,这些并不利用神经网络的基本特征,如平行和模拟状态变量。 取代它们,它们模仿依赖计算力的神经网络,导致不可持续的能源消耗和相对较低的速度。 完全平行和模拟的硬件有望克服这些挑战,但模拟神经噪音及其传播的影响,即累积,威胁这种方法。这里,我们首次分析了由噪音非线性神经元组成的平行深线神经网络中噪音的传播。我们开发了对这两种网络的分析处理方法,以突出基本机制,并培训了后向传播的网络。我们发现,噪音积累一般是捆绑的,增加的网络层层不会使超出这一限度的噪音比噪音比率恶化。 最重要的是,噪音积累可以在神经振动功能时完全抑制,即威胁这种方法。在这里,我们首次分析由噪音神经振动网络组成的平行深层神经网络的传播情况,因此我们开发了一个更小的硬的网络框架。