A wide variety of orthographic coding schemes and models of visual word identification have been developed to account for masked priming data that provide a measure of orthographic similarity between letter strings. These models tend to include hand-coded orthographic representations with single unit coding for specific forms of knowledge (e.g., units coding for a letter in a given position or a letter sequence). Here we assess how well a range of these coding schemes and models account for the pattern of form priming effects taken from the Form Priming Project and compare these findings to results observed in with 11 standard deep neural network models (DNNs) developed in computer science. We find that deep convolutional networks perform as well or better than the coding schemes and word recognition models, whereas transformer networks did less well. The success of convolutional networks is remarkable as their architectures were not developed to support word recognition (they were designed to perform well on object recognition) and they classify pixel images of words (rather artificial encodings of letter strings). The findings add to the recent work of (Hannagan et al., 2021) suggesting that convolutional networks may capture key aspects of visual word identification.
翻译:已经开发了各种各样的拼写编码办法和视觉文字识别模型,以解析为字母字符串之间拼写相似度的度量度的隐蔽边际数据。这些模型往往包括手码拼写拼写图示,为特定知识形式(例如,在特定位置或字母序列中为字母编码单位或字母序列)提供单一单位编码(例如,在特定位置或字母序列中为字母编码单位)。在这里,我们评估了这些编码办法和模型如何很好地考虑到从“形式优化项目”中得出的形式边际效应模式,并将这些结果与在计算机科学中开发的11个标准的深神经网络模型(DNNN)中观察到的结果进行比较。我们发现,深相联网络的运行或好于编码计划和字典识别模型,而变压器网络则不那么好。 进动网络的成功是惊人的,因为其结构没有发展起来支持文字识别(其设计目的是在对象识别上很好地执行),并将单词识别图象图像图像图像图像(而不是字母字符串的人工编码)。结论是对最近的工作(Hangan et al. 2021) 关键方面的建议。