Recently, a number of articles have argued that deep learning models such as GPT could also capture key aspects of language processing in the human mind and brain. However, I will argue that these models are not suitable as neural models of human language. Firstly, because they fail on fundamental boundary conditions, such as the amount of learning they require. This would in fact imply that the mechanisms of GPT and brain language processing are fundamentally different. Secondly, because they do not possess the logistics of access needed for compositional and productive human language processing. Neural architectures could possess logistics of access based on small-world like network structures, in which processing does not consist of symbol manipulation but of controlling the flow of activation. In this view, two complementary approaches would be needed to investigate the relation between brain and cognition. Investigating learning methods could reveal how 'learned cognition' as found in deep learning could develop in the brain. However, neural architectures with logistics of access should also be developed to account for 'productive cognition' as required for natural or artificial human language processing. Later on, these approaches could perhaps be combined to see how such architectures could develop by learning and development from a simpler basis.
翻译:最近,一些文章指出,GPT等深层次学习模式也可以捕捉到人脑和脑中语言处理的关键方面。然而,我将争辩说,这些模式不适宜作为人类语言的神经模型。首先,因为它们在基本边界条件下失败,例如学习所需的数量。这实际上意味着GPT和大脑语言处理机制根本不同。第二,因为它们不具备进入人类语言组成和生产性处理所需的后勤条件。神经结构可以拥有基于小世界(例如网络结构)的访问物流,而网络结构的处理并不包括符号操纵,而是控制激活的流动。在这种观点中,需要两种互补的方法来调查大脑和认知之间的关系。调查学习方法可以揭示在深层学习中发现的“认知”如何在大脑中发展。然而,与接触物流相关的神经结构也应该发展,以核算自然或人工人类语言处理所需要的“生产性认知”。随后,这些方法也许可以结合到如何通过从简单的基础学习和发展来发展这些结构。