The representations generated by many models of language (word embeddings, recurrent neural networks and transformers) correlate to brain activity recorded while people read. However, these decoding results are usually based on the brain's reaction to syntactically and semantically sound language stimuli. In this study, we asked: how does an LSTM (long short term memory) language model, trained (by and large) on semantically and syntactically intact language, represent a language sample with degraded semantic or syntactic information? Does the LSTM representation still resemble the brain's reaction? We found that, even for some kinds of nonsensical language, there is a statistically significant relationship between the brain's activity and the representations of an LSTM. This indicates that, at least in some instances, LSTMs and the human brain handle nonsensical data similarly.
翻译:许多语言模型(语言嵌入、经常性神经网络和变压器)产生的表达方式与人们阅读时记录到的大脑活动相关。然而,这些解码结果通常基于大脑对合成和语义健全的语言刺激的反应。在这个研究中,我们问道:LSTM(长期内存)语言模型(经(按大小)培训的关于语义和整体完整语言的(短期)语言模型如何代表退化的语义或合成信息的语言样本?LSTM(LSTM)代表着退化的语义或合成信息吗?LSTM(LSTM)代表着大脑的反应吗?我们发现,即使对于某些非感官语言来说,大脑活动与LSTM(LSTM)的表达方式之间也存在着具有统计意义的关系。这表明,至少在一些情形下,LSTMs(LSTMs)和人类大脑处理着类似的非感官数据。