The capabilities and limitations of BERT and similar models are still unclear when it comes to learning syntactic abstractions, in particular across languages. In this paper, we use the task of subordinate-clause detection within and across languages to probe these properties. We show that this task is deceptively simple, with easy gains offset by a long tail of harder cases, and that BERT's zero-shot performance is dominated by word-order effects, mirroring the SVO/VSO/SOV typology.
翻译:在学习综合抽象学方面,特别是语言之间的综合抽象学方面,BERT和类似模型的能力和局限性仍然不明确。在本文中,我们使用语言内部和语言之间的从属探测任务来探测这些属性。我们表明,这项任务是欺骗性的简单,容易得到的收益被长尾较难的案例所抵消,而BERT的零发射性能主要是单词顺序效果,这反映了SVO/VSO/SOV类型。