Circuits of biological neurons, such as in the functional parts of the brain can be modeled as networks of coupled oscillators. Inspired by the ability of these systems to express a rich set of outputs while keeping (gradients of) state variables bounded, we propose a novel architecture for recurrent neural networks. Our proposed RNN is based on a time-discretization of a system of second-order ordinary differential equations, modeling networks of controlled nonlinear oscillators. We prove precise bounds on the gradients of the hidden states, leading to the mitigation of the exploding and vanishing gradient problem for this RNN. Experiments show that the proposed RNN is comparable in performance to the state of the art on a variety of benchmarks, demonstrating the potential of this architecture to provide stable and accurate RNNs for processing complex sequential data.
翻译:生物神经的电路,例如大脑功能部分的生物神经电路,可以作为混合振动器网络的模型。受这些系统在保持状态变量约束的同时表达大量产出的能力的启发,我们为经常性神经网络提出了一个新的结构。我们提议的RNN基于对二级普通差异方程系统的时间分解,对受控非线性振动器网络的模型。我们证明了隐藏状态梯度的精确界限,从而缓解了这个RNN的爆炸和消散梯度问题。实验表明,拟议的RNN在各种基准方面的性能与最新水平相当,显示了这一结构为处理复杂的连续数据提供稳定和准确的RNNS的潜力。