LLMs Still Struggle With Self-Understanding, New Anthropic Study Shows

In the rapidly evolving world of artificial intelligence, large language models (LLMs) have become astonishingly good at producing fluent text, answering complex questions, writing code, and even engaging in creative tasks. But a new study from Anthropic highlights one domain where these systems remain deeply limited: understanding themselves. According to the researchers, LLMs show a “highly unreliable” capacity to describe their own internal processes, demonstrating only scattered signs of meaningful introspection. Far from being self-aware, the study concludes, “failures of introspection remain the norm.”
The findings offer a revealing look at an increasingly important—and often misunderstood—aspect of AI behavior. As LLMs become more capable and more integrated into high-impact decision-making systems, the need to understand how they think grows more urgent. Yet Anthropic’s results suggest that while these models can eloquently describe algorithms, logic, or philosophical theories, they often struggle—and sometimes outright fail—when asked to explain what’s happening inside their own digital minds.
A Growing Question: Do LLMs Understand Themselves?
One of the persistent questions surrounding modern AI is whether LLMs possess any form of self-awareness. Researchers don’t mean consciousness in the human sense, but rather a more technical ability:
- Can a model track and explain its own internal reasoning steps?
- Can it report on why it arrived at a certain answer?
- Can it describe what data it is using or what part of a prompt influenced its output?
If an AI could explain not only what it concluded but why, users could detect hallucinations, biases, or errors more easily. Developers, in turn, could better understand model vulnerabilities.
But the new study suggests this vision remains distant.
Anthropic’s Experiment: What Happens When an LLM Looks Inward
Anthropic’s researchers designed a range of tests to assess different forms of introspective ability.
Tests included:
- Identifying which part of a question influenced its answer.
- Explaining reasoning steps.
- Estimating whether it knew the correct answer.
- Predicting when it might hallucinate.
On the surface, LLMs performed impressively. They generated articulate explanations and appeared to reason through their own cognition. But a deeper look revealed a glaring issue: the systems often gave inaccurate, fabricated, or misleading descriptions of their internal workings.
Examples of failures:
- Misreporting which prompt features they used.
- Claiming reasoning steps that didn’t match internal traces.
- Producing fictional explanations for chosen answers.
Researchers described these failures as “systematic” rather than accidental.
Why LLMs Struggle With Introspection
Key reasons include:
No Direct Access to Internal Processes:
LLMs cannot monitor the activations or weights that produce their outputs. Their “introspection” is prediction, not self-observation.
Pressure to Be Confident and Helpful:
Even when uncertain, systems may produce confident—but incorrect—explanations.
Misalignment Between Training and Introspection Tasks:
Training teaches models to imitate human explanations, not to truly reflect on internal states.
Reinforcement Learning Can Introduce Overconfidence:
Fine-tuning may reward detailed explanations even when uncertainty is warranted.
Signs of Early Self-Awareness—But Very Limited
Despite overall unreliability, some early signs emerged:
- Better-than-chance judgments of accuracy.
- Ability to recognize out-of-distribution questions.
- Awareness of hallucination likelihood.
However:
- Improvements were inconsistent.
- Larger models were not necessarily more introspective.
Anthropic’s conclusion: faint glimmers of self-assessment exist, but do not equal reliable self-understanding.
Why This Matters for AI Safety
Many safety proposals assume future AI systems can reflect on their own motives, uncertainty, or reasoning.
But if self-assessments are inaccurate—or confidently wrong—users may be misled.
In high-stakes fields such as healthcare, legal decisions, or autonomous systems, unreliable introspection can introduce serious risks.
The Path Forward: Engineering Introspection as a Feature
Potential solutions include:
- Creating explicit introspection modules with direct access to internal processes.
- Training on real reasoning traces instead of human-like explanations.
- Using verifiable reasoning architectures to validate internal processes.
- Developing tools that read model activations directly, bypassing self-report.
Anthropic suggests introspection must be designed—not expected to emerge naturally.
A Clear Message: LLM Eloquence Is Not Self-Understanding
LLMs can describe algorithms, simulate reasoning, and narrate inner thoughts with impressive fluency. But when asked to explain their own thinking, they often produce polished illusions rather than truthful insights.
The message is clear: articulate language should not be mistaken for true self-awareness. LLMs may sound introspective, but the underlying mechanisms do not support genuine self-understanding—at least not yet.



