This work is intended as a voice in the discussion over the recent claims that LaMDA, a pretrained language model based on the Transformer model architecture, is sentient. This claim, if confirmed, would have serious ramifications in the Natural Language Processing (NLP) community due to wide-spread use of similar models. However, here we take the position that such a language model cannot be sentient, or conscious, and that LaMDA in particular exhibits no advances over other similar models that would qualify it. We justify this by analysing the Transformer architecture through Integrated Information Theory. We see the claims of consciousness as part of a wider tendency to use anthropomorphic language in NLP reporting. Regardless of the veracity of the claims, we consider this an opportune moment to take stock of progress in language modelling and consider the ethical implications of the task. In order to make this work helpful for readers outside the NLP community, we also present the necessary background in language modelling.
翻译:这项工作旨在为讨论最近关于LAMDA这一以变异模型结构为基础的预先培训语言模型具有感性的说法提供声音,如果这一说法得到证实,由于广泛使用类似的模型,会对自然语言处理(NLP)社区产生严重影响。然而,我们在这里采取的立场是,这种语言模型不能具有感知性或意识,特别是,LAMDA没有在其他类似模型上取得任何进展,因而符合资格。我们通过综合信息理论分析变异结构来证明这一点。我们认为,在NLP报告中使用人类形态语言的更广泛趋势中,意识的主张是其中的一部分。不管这些主张的真实性,我们认为这是评估语言建模方面的进展和考虑任务的道德影响的一个适当时机。为了使这项工作对NLP社区以外的读者有所帮助,我们还在语言建模中介绍了必要的背景。