This paper proposes a new architecture - Attentive Tensor Product Learning (ATPL) - to represent grammatical structures in deep learning models. ATPL is a new architecture to bridge this gap by exploiting Tensor Product Representations (TPR), a structured neural-symbolic model developed in cognitive science, aiming to integrate deep learning with explicit language structures and rules. The key ideas of ATPL are: 1) unsupervised learning of role-unbinding vectors of words via TPR-based deep neural network; 2) employing attention modules to compute TPR; and 3) integration of TPR with typical deep learning architectures including Long Short-Term Memory (LSTM) and Feedforward Neural Network (FFNN). The novelty of our approach lies in its ability to extract the grammatical structure of a sentence by using role-unbinding vectors, which are obtained in an unsupervised manner. This ATPL approach is applied to 1) image captioning, 2) part of speech (POS) tagging, and 3) constituency parsing of a sentence. Experimental results demonstrate the effectiveness of the proposed approach.
翻译:本文提出了一个新的结构――强化指标产品学习(ATPL)――在深层学习模型中代表语法结构结构结构结构结构结构结构。ATPL是一个通过利用Tonsor产品表征(TPR)弥合这一差距的新结构,TPR是一种在认知科学中开发的结构性神经-同步模型,旨在将深层次学习与明确的语言结构和规则相结合。ATPL的主要思想是:(1) 通过基于TRP的深层神经网络,不受监督地学习无约束的文字矢量;(2) 利用关注模块来计算TRP;(3) 将TRP与典型的深层学习结构相结合,包括长期短期内存和进取型神经网络。我们的方法的新颖之处在于它能够通过使用不强制的无约束矢量获取出一个句子的语法结构。这种ATPL方法适用于:(1) 图像说明,(2) 部分语音标记,和(3) 语系对句进行分解。实验结果表明拟议方法的有效性。