Deanthropomorphising NLP: Can a language model be conscious?

This work is intended as a voice in the discussion over previous claims that a pretrained large language model (LLM) based on the Transformer model architecture can be sentient. Such claims have been made concerning the LaMDA model and also concerning the current wave of LLM-powered chatbots, such a...

Full description

Saved in:
Bibliographic Details
Main Authors: Matthew Shardlow, Piotr Przybyła
Format: Article
Language:English
Published: Public Library of Science (PLoS) 2024-01-01
Series:PLoS ONE
Online Access:https://doi.org/10.1371/journal.pone.0307521
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1850106369683750912
author Matthew Shardlow
Piotr Przybyła
author_facet Matthew Shardlow
Piotr Przybyła
author_sort Matthew Shardlow
collection DOAJ
description This work is intended as a voice in the discussion over previous claims that a pretrained large language model (LLM) based on the Transformer model architecture can be sentient. Such claims have been made concerning the LaMDA model and also concerning the current wave of LLM-powered chatbots, such as ChatGPT. This claim, if confirmed, would have serious ramifications in the Natural Language Processing (NLP) community due to wide-spread use of similar models. However, here we take the position that such a large language model cannot be conscious, and that LaMDA in particular exhibits no advances over other similar models that would qualify it. We justify this by analysing the Transformer architecture through Integrated Information Theory of consciousness. We see the claims of sentience as part of a wider tendency to use anthropomorphic language in NLP reporting. Regardless of the veracity of the claims, we consider this an opportune moment to take stock of progress in language modelling and consider the ethical implications of the task. In order to make this work helpful for readers outside the NLP community, we also present the necessary background in language modelling.
format Article
id doaj-art-0f37ff4fbe034b3e9f3e4fbe57da3cf1
institution OA Journals
issn 1932-6203
language English
publishDate 2024-01-01
publisher Public Library of Science (PLoS)
record_format Article
series PLoS ONE
spelling doaj-art-0f37ff4fbe034b3e9f3e4fbe57da3cf12025-08-20T02:38:51ZengPublic Library of Science (PLoS)PLoS ONE1932-62032024-01-011912e030752110.1371/journal.pone.0307521Deanthropomorphising NLP: Can a language model be conscious?Matthew ShardlowPiotr PrzybyłaThis work is intended as a voice in the discussion over previous claims that a pretrained large language model (LLM) based on the Transformer model architecture can be sentient. Such claims have been made concerning the LaMDA model and also concerning the current wave of LLM-powered chatbots, such as ChatGPT. This claim, if confirmed, would have serious ramifications in the Natural Language Processing (NLP) community due to wide-spread use of similar models. However, here we take the position that such a large language model cannot be conscious, and that LaMDA in particular exhibits no advances over other similar models that would qualify it. We justify this by analysing the Transformer architecture through Integrated Information Theory of consciousness. We see the claims of sentience as part of a wider tendency to use anthropomorphic language in NLP reporting. Regardless of the veracity of the claims, we consider this an opportune moment to take stock of progress in language modelling and consider the ethical implications of the task. In order to make this work helpful for readers outside the NLP community, we also present the necessary background in language modelling.https://doi.org/10.1371/journal.pone.0307521
spellingShingle Matthew Shardlow
Piotr Przybyła
Deanthropomorphising NLP: Can a language model be conscious?
PLoS ONE
title Deanthropomorphising NLP: Can a language model be conscious?
title_full Deanthropomorphising NLP: Can a language model be conscious?
title_fullStr Deanthropomorphising NLP: Can a language model be conscious?
title_full_unstemmed Deanthropomorphising NLP: Can a language model be conscious?
title_short Deanthropomorphising NLP: Can a language model be conscious?
title_sort deanthropomorphising nlp can a language model be conscious
url https://doi.org/10.1371/journal.pone.0307521
work_keys_str_mv AT matthewshardlow deanthropomorphisingnlpcanalanguagemodelbeconscious
AT piotrprzybyła deanthropomorphisingnlpcanalanguagemodelbeconscious