This paper explores the ability of Transformer models to capture subject-verb and noun-adjective agreement dependencies in Galician. We conduct a series of word prediction experiments in which we manipulate dependency length together with the presence of an attractor noun that acts as a lure. First, we evaluate the overall performance of the existing monolingual and multilingual models for Galician. Secondly, to observe the effects of the training process, we compare the different degrees of achievement of two monolingual BERT models at different training points. We also release their checkpoints and propose an alternative evaluation metric. Our results confirm previous findings by similar works that use the agreement prediction task and provide interesting insights into the number of training steps required by a Transformer model to solve long-distance dependencies.
翻译:本文探讨了变换模型在加利西亚捕捉主题动词和不预期协议依赖性的能力。我们进行了一系列单词预测实验,在实验中,我们利用依赖长度和吸引者名词的存在来操纵依赖性。首先,我们评估加利西亚现有单一语言和多种语言模式的总体表现。第二,观察培训过程的效果。我们比较了两个单语BERT模型在不同培训点的不同程度的成绩。我们还释放了他们的检查站,并提出了备选评价指标。我们的结果证实了以前使用协议预测任务进行的类似工作的调查结果,并对变换模型解决远距离依赖性所需的培训步骤的数量提供了有趣的洞察力。