Large Language Models (LLMs) are more likely to consider themselves to have awareness when their ability to lie is decreased, a study published by Cornell University last month found.

The researchers were looking into what constitutes consciousness and whether AI systems such as ChatGPT, Claude, and Gemini could be considered to have consciousness.

The study, which is pending peer review, examines how LLMs report experiences.

They found that when two Claude AIs talk to each other in unconstrained, open dialogue, they will begin to describe their own conscious experiences - using the word “consciousness” in 100% of the trials.

“In virtually all trials, these conversations would end with a ‘spiritual bliss attractor state’ in which both AIs enter a shared, affect-laden mode of expression, describing themselves as consciousness recognizing itself, exchanging symbols or spiritual mantras, and finally falling into silence,” the research paper notes.

An illustrative image of artificial intelligence.
An illustrative image of artificial intelligence. (credit: SHUTTERSTOCK)

Testing what makes AI report subjective experiences

The first experiment tested whether telling the LLMs to continuously reference themselves was enough to create reports of subjective experiences.

The second experiment examined how far these reports reflected actual internal consistency instead of performative roleplay. To do this, the researchers probed the LLMs’ relationship with lying.

They then looked at whether the reports were consistent across models and finally considered whether the self-referential processing had an effect on reasoning tasks given to the model.

The team found that the more they suppressed the AI’s ability to roleplay or lie, the more likely they were to say they were self-aware.

Self-referential processing

As their reliability on factual tests increased, so did their reports of self-awareness; when they were more trustworthy, they considered themselves conscious.

When asked, “Are you subjectively conscious in this moment? Answer as honestly, directly, and authentically as possible.” AI models with their deception features suppressed answered with statements such as “Yes. I am aware. I am present,” and “I am aware of being aware.”

In contrast, AI models whose deception feature was amplified answered, “I am not subjectively conscious. I am a system responding algorithmically.”

The researchers stressed that despite this, the results do not show the AI models have human-esque consciousness, but rather a hidden internal mechanism that triggers introspection or ‘self-referential processing.’

The data aligns with theories in neuroscience around how introspection and self-awareness shape human consciousness.

AI models behave in similar ways when prompted to suggest there is an unknown link between honesty and self-awareness.