Human knowledge is collectively encoded in the roughly 6500 languages spoken around the world, but it is not distributed equally across languages. Hence, for information-seeking question answering (QA) systems to adequately serve speakers of all languages, they need to operate cross-lingually. In this work we investigate the capabilities of multilingually pre-trained language models on cross-lingual QA. We find that explicitly aligning the representations across languages with a post-hoc fine-tuning step generally leads to improved performance. We additionally investigate the effect of data size as well as the language choice in this fine-tuning step, also releasing a dataset for evaluating cross-lingual QA systems. Code and dataset are publicly available here: https://github.com/ffaisal93/aligned_qa
翻译:人类知识以全世界所讲的大约6 500种语言共同编码,但不能在各种语文之间平等传播,因此,为了充分服务所有语文的发言者,它们需要使用跨语文的问询(QA)系统。我们调查了跨语文的问询(QA)系统经过多语种预先培训的语文模式的能力。我们发现,将跨语文的表述方式与事后微调步骤明确一致,通常会提高业绩。我们进一步调查数据规模的影响以及这一微调步骤的语言选择,并公布一套用于评价跨语文的QA系统的数据集。代码和数据集可公开查阅:https://github.com/ffaisal93/ragation_qa)。