Natural Language Processing (NLP) has become one of the leading application areas in the current Artificial Intelligence boom. Transfer learning has enabled large deep learning neural networks trained on the language modeling task to vastly improve performance in almost all language tasks. Interestingly, when the models are trained with data that includes software code, they demonstrate remarkable abilities in generating functioning computer code from natural language specifications. We argue that this creates a conundrum for claims that neural models provide an alternative theory to generative phrase structure grammars in explaining how language works. Since the syntax of programming languages is determined by phrase structure grammars, successful neural models are apparently uninformative about the theoretical foundations of programming languages, and by extension, natural languages. We argue that the term language model is misleading because deep learning models are not theoretical models of language and propose the adoption of corpus model instead, which better reflects the genesis and contents of the model.
翻译:自然语言处理(NLP)已成为当前人工智能发展的主要应用领域之一; 转移学习使在语言模型任务方面受过培训的大型深层次学习神经网络得以大大改进几乎所有语言任务的业绩。有趣的是,当模型经过包括软件代码在内的数据培训时,它们展示了根据自然语言规格生成功能计算机代码的非凡能力。我们争辩说,这为神经模型在解释语言如何运作方面提供了变异的词组结构语法结构的替代理论这一说法制造了一个难题。 由于编程语言的语法由语法结构的短语决定,成功的神经模型显然对编程语言的理论基础以及扩展的自然语言缺乏信息。我们争辩说,术语模式具有误导性,因为深层次学习模型不是语言的理论模型,因此建议采用文理模型,从而更好地反映该模型的起源和内容。