In 2018, Mikolov et al. introduced the positional language model, which has characteristics of attention-based neural machine translation models and which achieved state-of-the-art performance on the intrinsic word analogy task. However, the positional model is not practically fast and it has never been evaluated on qualitative criteria or extrinsic tasks. We propose a constrained positional model, which adapts the sparse attention mechanism from neural machine translation to improve the speed of the positional model. We evaluate the positional and constrained positional models on three novel qualitative criteria and on language modeling. We show that the positional and constrained positional models contain interpretable information about the grammatical properties of words and outperform other shallow models on language modeling. We also show that our constrained model outperforms the positional model on language modeling and trains twice as fast.
翻译:2018年,Mikolov等人引入了定位语言模型,该模型具有以关注为基础的神经机器翻译模型的特点,在内在的单词类比任务上达到了最先进的性能。然而,定位模型实际上并不快,而且从未对质量标准或外表任务进行过评估。我们提出了一个受制约的定位模型,该模型将稀疏的注意力机制从神经机翻译中调整,以提高定位模型的速度。我们评估了三种新颖质量标准和语言模型的定位和受限位置模型。我们显示,定位和受限位置模型包含关于语言语法特性的可解释信息,并超越了其他语言模型的外形模型。我们还表明,我们受限制的模型比语言模型的定位模型和训练速度高出一倍。