Recent work has demonstrated that machine learning models allow us to compare languages by showing how hard each language might be to learn under specific tasks. Following this line of investigation, we investigate what makes a language "hard to pronounce" by modelling the task of grapheme-to-phoneme (g2p) transliteration. By training a character-level transformer model on this task across 22 languages and measuring the model's proficiency against its grapheme and phoneme inventories, we show that certain characteristics emerge that separate easier and harder languages with respect to learning to pronounce. Namely that the complexity of a languages pronunciation from its orthography is due to how expressive or simple its grapheme-to-phoneme mapping is. Further discussion illustrates how future studies should consider relative data sparsity per language in order to design more fair cross lingual comparison tasks.
翻译:最近的工作表明,机器学习模式通过显示每种语言在具体任务下学习的难度,使我们能够比较语言。根据这一调查方针,我们通过模拟石墨到手机(g2p)的音异化任务来调查是什么使得一种语言“很难发音 ” 。 通过在22种语言中培训一个关于这项任务的品格级变压器模型,并对照其石墨和电话目录来测量模型的熟练程度,我们表明,在学习读写时,出现了一些比较容易和较难的语言。也就是说,一种语言的发音的复杂性是由于其图形到语音的绘图如何表达或简单。进一步的讨论表明,未来的研究应该如何考虑每种语言相对的数据宽度,以便设计更公平的跨语言比较任务。