Trans-dimensional random field language models (TRF LMs) have recently been introduced, where sentences are modeled as a collection of random fields. The TRF approach has been shown to have the advantages of being computationally more efficient in inference than LSTM LMs with close performance and being able to flexibly integrating rich features. In this paper we propose neural TRFs, beyond of the previous discrete TRFs that only use linear potentials with discrete features. The idea is to use nonlinear potentials with continuous features, implemented by neural networks (NNs), in the TRF framework. Neural TRFs combine the advantages of both NNs and TRFs. The benefits of word embedding, nonlinear feature learning and larger context modeling are inherited from the use of NNs. At the same time, the strength of efficient inference by avoiding expensive softmax is preserved. A number of technical contributions, including employing deep convolutional neural networks (CNNs) to define the potentials and incorporating the joint stochastic approximation (JSA) strategy in the training algorithm, are developed in this work, which enable us to successfully train neural TRF LMs. Various LMs are evaluated in terms of speech recognition WERs by rescoring the 1000-best lists of WSJ'92 test data. The results show that neural TRF LMs not only improve over discrete TRF LMs, but also perform slightly better than LSTM LMs with only one fifth of parameters and 16x faster inference efficiency.
翻译:最近引入了跨度随机外野语言模型(TRF LM),该模型的句号以随机字段的集合为模型,显示TRF方法的优点是,在计算上比LSTM LMLM更高效的推论中,性能接近,并能灵活整合丰富的特点。在本文件中,我们提议神经TRF,除了前一个仅使用具有离散特性的线性潜力的离散性TRF之外,我们建议神经TRF, 其想法是利用神经网络在TRF框架内实施的具有连续特征的非线性潜力。神经TRF将NS和TRF的优势结合起来。字嵌嵌入、非线性特征学习和更大背景模型的优点来自NNS的使用。与此同时,我们建议保持通过避免昂贵软体软体的离散性TRF格式而有效推论的强度。一些技术贡献,包括使用神经神经网络(CNNS)来界定潜力和将联合神经近MS(JSA)的参数结合NFS和TRMLMS的优势。 在培训算中,只对LRRF的某种语言测试结果进行了一项成功的评估,这使我们得以成功地认识。