Despite the rapid recent progress in creating accurate and compact in-context learners, most recent work focuses on in-context learning (ICL) for tasks in English. However, the ability to interact with users of languages outside English presents a great potential for broadening the applicability of language technologies to non-English speakers. In this work, we collect the infrastructure necessary for training and evaluation of ICL in a selection of Slavic languages: Czech, Polish, and Russian. We link a diverse set of datasets and cast these into a unified instructional format through a set of transformations and newly-crafted templates written purely in target languages. Using the newly-curated dataset, we evaluate a set of the most recent in-context learners and compare their results to the supervised baselines. Finally, we train, evaluate and publish a set of in-context learning models that we train on the collected resources and compare their performance to previous work. We find that ICL models tuned in English are also able to learn some tasks from non-English contexts, but multilingual instruction fine-tuning consistently improves the ICL ability. We also find that the massive multitask training can be outperformed by single-task training in the target language, uncovering the potential for specializing in-context learners to the language(s) of their application.
翻译:翻译摘要:
尽管最近在创建准确且紧凑的上下文学习器方面取得了快速进展,但大部分最新研究都集中在针对英语任务的上下文学习(ICL)上。然而,与非英语母语用户互动的能力为将语言技术应用范围扩展到非英语使用者提供了巨大潜力。在这项工作中,我们收集Slavic语言选择的ICL的培训和评估基础设施:捷克语、波兰语和俄语。我们通过一系列转换和全新的纯目标语编写的模板,将多样化的数据集连接成一个统一的教学格式。使用新策划的数据集,我们评估了一组最新的上下文学习器,并将其结果与监督基准进行了比较。最后,我们培训、评估和发布一组上下文学习模型,这些模型使用收集的资源进行培训,并将其性能与之前的研究进行比较。我们发现,英文调优的ICL模型也能够从非英语环境中学习一些任务,但多语言指导微调始终提高了ICL的能力。我们还发现,对单语言的单一任务培训可能比对目标语言的各种任务进行的大规模多任务培训效果更好,揭示了让上下文学习器专门化于其应用语言的潜力。