To what extent can neural network models learn generalizations about language structure, and how do we find out what they have learned? We explore these questions by training neural models for a range of natural language processing tasks on a massively multilingual dataset of Bible translations in 1295 languages. The learned language representations are then compared to existing typological databases as well as to a novel set of quantitative syntactic and morphological features obtained through annotation projection. We conclude that some generalizations are surprisingly close to traditional features from linguistic typology, but that most of our models, as well as those of previous work, do not appear to have made linguistically meaningful generalizations. Careful attention to details in the evaluation turns out to be essential to avoid false positives. Furthermore, to encourage continued work in this field, we release several resources covering most or all of the languages in our data: (i) multiple sets of language representations, (ii) multilingual word embeddings, (iii) projected and predicted syntactic and morphological features, (iv) software to provide linguistically sound evaluations of language representations.
翻译:神经网络模型能够在多大程度上了解对语言结构的概括,以及我们如何发现它们所学到的?我们通过培训神经模型来探讨这些问题,在1295种语言的圣经翻译大量多语文数据集方面,为一系列自然语言处理任务培训神经模型;然后,将所学的语言表述方式与现有的类型数据库以及通过注解预测获得的一套新型定量综合和形态特征进行比较;我们的结论是,有些概括方式与语言类型的传统特征格外接近,但大多数我们的模式以及以前的工作模式似乎没有产生具有语言意义的概括性。仔细注意评价中的细节对于避免错误的正面效果至关重要。此外,为了鼓励继续这一领域的工作,我们将数据中涵盖大多数或所有语言的若干资源放入我们的数据:(一)多套语言表述方式,(二)多语言嵌入语言,(三)预测和预测的合成和形态特征,(四)提供语言正确评价语言表达方式的软件。