Long Short-Term Memory (LSTM) Recurrent Neural networks (RNNs) rely on gating signals, each driven by a function of a weighted sum of at least 3 components: (i) one of an adaptive weight matrix multiplied by the incoming external input vector sequence, (ii) one adaptive weight matrix multiplied by the previous memory/state vector, and (iii) one adaptive bias vector. In effect, they augment the simple Recurrent Neural Networks (sRNNs) structure with the addition of a "memory cell" and the incorporation of at most 3 gating signals. The standard LSTM structure and components encompass redundancy and overly increased parameterization. In this paper, we systemically introduce variants of the LSTM RNNs, referred to as SLIM LSTMs. These variants express aggressively reduced parameterizations to achieve computational saving and/or speedup in (training) performance---while necessarily retaining (validation accuracy) performance comparable to the standard LSTM RNN.
翻译:长期内存(LSTM) 常规神经网络(RNNS) 依赖光学信号,每个信号的驱动力是至少3个组成部分的加权总和的函数:(一) 一个适应性加权矩阵乘以外来输入矢量序列,(二) 一个适应性加权矩阵乘以以前的内存/状态矢量,(三) 一个适应性偏向矢量,实际上,它们增加了简单的经常性神经网络结构,增加了一个“模拟细胞”,最多包括3个标记信号。标准的LSTM结构和组成部分包括冗余和过度增加参数化。在本文件中,我们系统地采用了称为SLIM LSTMS的LSTMRNNS变体。这些变体表示急剧减少参数化,以便在(培训)性能(性能)中实现计算储蓄和(或)加速,同时必须保留与标准LSTM RNNSN(价值精度) 的性能。