Sequential neural networks models are powerful tools in a variety of Natural Language Processing (NLP) tasks. The sequential nature of these models raises the questions: to what extent can these models implicitly learn hierarchical structures typical to human language, and what kind of grammatical phenomena can they acquire? We focus on the task of agreement prediction in Basque, as a case study for a task that requires implicit understanding of sentence structure and the acquisition of a complex but consistent morphological system. Analyzing experimental results from two syntactic prediction tasks -- verb number prediction and suffix recovery -- we find that sequential models perform worse on agreement prediction in Basque than one might expect on the basis of a previous agreement prediction work in English. Tentative findings based on diagnostic classifiers suggest the network makes use of local heuristics as a proxy for the hierarchical structure of the sentence. We propose the Basque agreement prediction task as challenging benchmark for models that attempt to learn regularities in human language.
翻译:序列神经网络模型是各种自然语言处理(NLP)任务中的有力工具。这些模型的顺序性质提出了问题:这些模型在多大程度上可以隐含地学习人类语言典型的等级结构,以及它们能够取得何种语法现象?我们侧重于巴斯克的协议预测任务,以此作为一项案例研究,以完成一项需要隐含地理解判刑结构和获得复杂但一致的形态系统的任务。分析两种综合预测任务 -- -- 动数预测和后缀恢复 -- -- 的实验结果 -- 我们发现,在巴斯克的协议预测方面,顺序模型在巴斯克的绩效比在先前的英语协议预测工作基础上人们可能预期的要差。基于诊断分类者的暂定调查结果表明,网络利用当地牛皮作为该句的等级结构的代用。我们建议,巴斯克协议预测任务是试图学习人类语言规律的模型具有挑战性的基准。