We show that both an LSTM and a unitary-evolution recurrent neural network (URN) can achieve encouraging accuracy on two types of syntactic patterns: context-free long distance agreement, and mildly context-sensitive cross serial dependencies. This work extends recent experiments on deeply nested context-free long distance dependencies, with similar results. URNs differ from LSTMs in that they avoid non-linear activation functions, and they apply matrix multiplication to word embeddings encoded as unitary matrices. This permits them to retain all information in the processing of an input string over arbitrary distances. It also causes them to satisfy strict compositionality. URNs constitute a significant advance in the search for explainable models in deep learning applied to NLP.
翻译:我们发现,LSTM和单一-革命经常性神经网络(URN)在两类综合模式上都能够取得令人鼓舞的准确性:无上下文长距离协议和对背景有轻微敏感认识的跨序列依赖。这项工作扩展了最近关于深嵌入无上下文长距离依赖性的实验,结果相似。URN与LSTMs不同,因为它们避免了非线性激活功能,并且对以单一矩阵编码的单词嵌入器应用矩阵乘法。这使得它们能够在任意距离的输入字符串处理中保留所有信息。这也导致它们满足严格的构成性。URN在深入学习适用于NLP的可解释模型方面迈出了一大进步。