Recently, there has been interest in multiplicative recurrent neural networks for language modeling. Indeed, simple Recurrent Neural Networks (RNNs) encounter difficulties recovering from past mistakes when generating sequences due to high correlation between hidden states. These challenges can be mitigated by integrating second-order terms in the hidden-state update. One such model, multiplicative Long Short-Term Memory (mLSTM) is particularly interesting in its original formulation because of the sharing of its second-order term, referred to as the intermediate state. We explore these architectural improvements by introducing new models and testing them on character-level language modeling tasks. This allows us to establish the relevance of shared parametrization in recurrent language modeling.
翻译:最近,人们对多种重复性经常性神经网络进行语言建模感兴趣。 事实上,简单的经常性神经网络(RNN)在生成隐藏状态之间高度关联的序列时遇到从过去错误中恢复过来的困难。 这些挑战可以通过在隐藏状态更新中整合二阶术语来缓解。 其中一个模型,即多倍性长短期内存(mLSTM),在最初的编制中特别有趣,因为它的第二阶术语被称为中间状态。 我们通过引入新模型和测试字符级语言建模任务来探索这些建筑改进。 这使我们能够在经常性语言建模中建立共同的对应关系。