Neural network models have been proposed to explain the grapheme-phoneme mapping process in humans for many alphabet languages. These models not only successfully learned the correspondence of the letter strings and their pronunciation, but also captured human behavior in nonce word naming tasks. How would the neural models perform for a non-alphabet language (e.g., Chinese) unknown character task? How well would the model capture human behavior? In this study, we evaluate a set of transformer models and compare their performances with human behaviors on an unknown Chinese character naming task. We found that the models and humans behaved very similarly, that they had similar accuracy distribution for each character, and had a substantial overlap in answers. In addition, the models' answers are highly correlated with humans' answers. These results suggested that the transformer models can well capture human's character naming behavior.
翻译:神经网络模型已被用来解释多种字母表语言中字素-音素映射过程,并且不仅成功地学习到了字母串与其发音的对应关系,还可以捕捉到人类的行为。但是这些模型在非字母表语言(如汉语)的未知字符任务中表现如何?它们如何捕捉到人类的行为?在本研究中,我们评估了一组变压器模型,并将其表现与人类行为在未知汉字命名任务中进行了比较。我们发现,这些模型和人类的行为非常相似,对于每一个字符的准确率分布也非常类似,并且答案有很大重叠。此外,模型的答案与人类的答案高度相关。这些结果表明,变压器模型可以很好地捕捉到人类的字符命名行为。