论文标题

Deanthropomerpormenting NLP:语言模型可以意识到吗?

Deanthropomorphising NLP: Can a Language Model Be Conscious?

论文作者

Shardlow, Matthew, Przybyła, Piotr

论文摘要

这项工作旨在作为对先前声称的讨论中的声音,即基于变压器模型体系结构的大型语言模型(LLM)可能是一种启发性的。已经提出了有关LAMDA模型的说法,还涉及当前的LLM驱动聊天机器人(例如Chatgpt)的浪潮。如果确认,由于广泛使用类似模型,该主张将在自然语言处理(NLP)社区中产生严重的影响。但是,在这里,我们认为这样的大语言模型不能是有意识的或有意识的,并且尤其是LAMDA对其他有资格的类似模型没有任何进展。我们通过通过综合信息理论分析变压器体系结构来证明这一点。我们认为,感知性的主张是在NLP报告中使用拟人化语言的更广泛倾向的一部分。无论主张的真实性如何,我们都认为这是一个合适的时机,可以盘点语言建模的进步并考虑任务的道德含义。为了使这项工作对NLP社区以外的读者有所帮助,我们还提供了语言建模的必要背景。

This work is intended as a voice in the discussion over previous claims that a pretrained large language model (LLM) based on the Transformer model architecture can be sentient. Such claims have been made concerning the LaMDA model and also concerning the current wave of LLM-powered chatbots, such as ChatGPT. This claim, if confirmed, would have serious ramifications in the Natural Language Processing (NLP) community due to wide-spread use of similar models. However, here we take the position that such a large language model cannot be sentient, or conscious, and that LaMDA in particular exhibits no advances over other similar models that would qualify it. We justify this by analysing the Transformer architecture through Integrated Information Theory of consciousness. We see the claims of sentience as part of a wider tendency to use anthropomorphic language in NLP reporting. Regardless of the veracity of the claims, we consider this an opportune moment to take stock of progress in language modelling and consider the ethical implications of the task. In order to make this work helpful for readers outside the NLP community, we also present the necessary background in language modelling.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源