A wide variety of orthographic coding schemes and models of visual word identification have been developed to account for masked priming data that provide a measure of orthographic similarity between letter strings. These models tend to include hand-coded orthographic representations with single unit coding for specific forms of knowledge (e.g., units coding for a letter in a given position). Here we assess how well a range of these coding schemes and models account for the pattern of form priming effects taken from the Form Priming Project and compare these findings to results observed with 11 standard deep neural network models (DNNs) developed in computer science. We find that deep convolutional networks (CNNs) perform as well or better than the coding schemes and word recognition models, whereas transformer networks did less well. The success of CNNs is remarkable as their architectures were not developed to support word recognition (they were designed to perform well on object recognition), they classify pixel images of words (rather than artificial encodings of letter strings), and their training was highly simplified (not respecting many key aspects of human experience). In addition to these form priming effects, we find that the DNNs can account for visual similarity effects on priming that are beyond all current psychological models of priming. The findings add to the recent work of (Hannagan et al., 2021) and suggest that CNNs should be given more attention in psychology as models of human visual word recognition.
翻译:已经开发了各种各样的拼写编码办法和视觉字辨识模型,以解算为字母字符串之间正拼写相似度的度量度的隐蔽直线网络模型所观察到的结果。这些模型往往包括手码正拼拼写图示,为特定形式的知识提供单一单位编码(例如,在特定位置为字母编码单位)。在这里,我们评估了这些编码办法和模型如何很好地考虑到从“表位原始项目”中得出的形式边缘效应模式,并将这些结果与在计算机科学中开发的11个标准的深层神经网络模型(DNNN)所观察到的结果进行比较。我们发现,深层相色网络(CNNN)与最近的编码计划和词辨识模型一样或更好,而变异网络则不那么好。 CNNM的架构之所以成功,是因为其结构没有发展起来支持文字识别(它们的设计是为了在对象识别上表现得当好)、对言词的像图像图像图像图像(而不是人为字母字符串的编码)进行分类,而且其培训非常简化(不尊重人类的很多关键直径直径网络网络模型方面) 。除了这些直观图象模型外,我们还可以图象学上所有的图象学模型还算外,还能够将所有的图象学影响。</s>