Natural Language Processing (NLP) has become one of the leading application areas in the current Artificial Intelligence boom. Transfer learning has enabled large deep learning neural networks trained on the language modeling task to vastly improve performance in almost all downstream language tasks. Interestingly, when the language models are trained with data that includes software code, they demonstrate remarkable abilities in generating functioning computer code from natural language specifications. We argue that this creates a conundrum for the claim that eliminative neural models are a radical restructuring in our understanding of cognition in that they eliminate the need for symbolic abstractions like generative phrase structure grammars. Because the syntax of programming languages is by design determined by phrase structure grammars, neural models that produce syntactic code are apparently uninformative about the theoretical foundations of programming languages. The demonstration that neural models perform well on tasks that involve clearly symbolic systems, proves that they cannot be used as an argument that language and other cognitive systems are not symbolic. Finally, we argue as a corollary that the term language model is misleading and propose the adoption of the working term corpus model instead, which better reflects the genesis and contents of the model.
翻译:自然语言处理(NLP)已成为当前人工智能繁荣中的主要应用领域之一。 传输学习使得在语言模型任务方面受过培训的大型深层次学习神经网络能够大大改进几乎所有下游语言任务的业绩。有趣的是,语言模型经过包括软件代码的数据培训,显示出在根据自然语言规格生成功能计算机代码方面的非凡能力。我们争辩说,这为以下说法制造了一个难题,即消除神经神经模型是我们对认知学的理解的彻底重组,因为它们消除了对象基因化短语结构语法学这样的象征性抽象学的需要。因为语言编程语言的合成是用短语结构语法来决定的,而产生合成代码的神经模型显然对编程语言的理论基础没有意义。神经模型在涉及明确象征性系统的任务上表现良好,证明它们不能用来作为语言和其他认知系统不是象征性的论据。最后,我们认为,语言模型的推论断是误导,并提议采用工作术语模型,从而更好地反映起源和内容。