Using Bayes's theorem, we derive a unit-wise recurrence as well as a backward recursion similar to the forward-backward algorithm. The resulting Bayesian recurrent units can be integrated as recurrent neural networks within deep learning frameworks, while retaining a probabilistic interpretation from the direct correspondence with hidden Markov models. Whilst the contribution is mainly theoretical, experiments on speech recognition indicate that adding the derived units at the end of state-of-the-art recurrent architectures can improve the performance at a very low cost in terms of trainable parameters.
翻译:使用 Bayes 的理论, 我们得出一个单位错误的重现以及类似于前向后向算法的后向重现。 由此形成的Bayesian 常规单位可以作为经常性神经网络纳入深层次学习框架,同时保留与隐蔽的Markov 模型直接对应的概率解释。 虽然其贡献主要是理论性的,但语音识别实验表明,在最先进的经常结构结束时添加衍生单位可以以非常低的成本提高可训练参数的性能。