Owing to the lack of corpora for low-resource languages, current works on dialogue generation have mainly focused on English. In this paper, we present mDIA, the first large-scale multilingual benchmark for dialogue generation across low- to high-resource languages. It covers real-life conversations in 46 languages across 19 language families. We present baseline results obtained by fine-tuning the multilingual, non-dialogue-focused pre-trained model mT5 as well as English-centric, dialogue-focused pre-trained chatbot DialoGPT. The results show that mT5-based models perform better on sacreBLEU and BertScore but worse on diversity. Even though promising results are found in few-shot and zero-shot scenarios, there is a large gap between the generation quality in English and other languages. We hope that the release of mDIA could encourage more works on multilingual dialogue generation to promote language diversity.
翻译:由于缺乏低资源语言的组合,目前关于对话的生成工作主要集中在英语上,本文介绍的是MDIA,这是在低资源语言和高资源语言之间开展对话的第一个大型多语文基准,涵盖19个语言家庭用46种语言进行的实际对话。我们介绍的是通过微调多语言、非对话重点的预先培训模式MT5以及以英语为中心的、以对话为重点的预先培训的聊天器DialoGPT所取得的基线结果。结果显示,基于MT5的模型在sacrebleU和BertScore多样性方面效果更好,但尽管在几率和零率的假设中都发现了有希望的结果,但英语和其他语言的一代质量之间仍然有很大差距。我们希望,发布MDIA能够鼓励更多关于多语言对话的生成工作,以促进语言多样性。