GPT-4o Lacks Core 'Theory of Mind' Features, Study Finds

New research suggests advanced LLMs don't truly understand mental states despite social proficiency.

A recent study challenges the notion that Large Language Models (LLMs) like GPT-4o possess a genuine 'Theory of Mind' (ToM). Researchers found that while LLMs can mimic human social judgments, they lack a consistent, domain-general understanding of how mental states drive behavior. This suggests their social abilities are more superficial than previously thought.

Katie Rowan

By Katie Rowan

February 15, 2026

4 min read

GPT-4o Lacks Core 'Theory of Mind' Features, Study Finds

Key Facts

  • A new study by John Muchovej and co-authors investigates GPT-4o's Theory of Mind (ToM).
  • The research found that GPT-4o lacks core features of ToM, despite its social proficiency.
  • LLMs succeed in approximating human judgments in simple ToM tasks but fail at logically equivalent ones.
  • The study highlights low consistency between GPT-4o's action predictions and mental state inferences.
  • The findings suggest LLM social abilities are not due to a domain-general or consistent ToM.

Why You Care

Ever wonder if the AI you chat with truly understands you? Can it really grasp your intentions or beliefs? A new study reveals that even models like GPT-4o might not have the deep social intelligence we imagine. This research suggests that while AI can mimic human conversation, its ‘understanding’ of mental states is still quite limited. Why should you care? Because it impacts how we interact with AI and what we expect from it in the future.

What Actually Happened

Researchers John Muchovej, Amanda Royka, Shane Lee, and Julian Jara-Ettinger published a paper titled “GPT-4o Lacks Core Features of Theory of Mind.” This study, submitted to CogSci 2025, investigates whether Large Language Models (LLMs) possess a Theory of Mind (ToM). ToM is the ability to attribute mental states—beliefs, intentions, desires—to oneself and others. The team developed a new evaluation structure, as detailed in the blog post. This structure probes if LLMs have a coherent model of how mental states cause behavior. They specifically GPT-4o, a prominent LLM. The findings indicate that while GPT-4o can approximate human judgments in simple ToM tasks, it struggles with logically equivalent scenarios. What’s more, it shows low consistency between its predictions of actions and its inferences about mental states, according to the announcement.

Why This Matters to You

This research has significant implications for how we perceive AI’s capabilities. If an AI doesn’t truly understand why someone acts a certain way, its ability to engage in complex social interactions is limited. Imagine you’re building an AI companion. If it lacks a genuine Theory of Mind, your AI might offer responses that seem human-like but are ultimately superficial. “We find that even though LLMs succeed in approximating human judgments in a simple ToM paradigm, they fail at a logically equivalent task and exhibit low consistency between their action predictions and corresponding mental state inferences,” the paper states. This suggests that the AI’s ‘social proficiency’ is not due to a deep understanding. What does this mean for your daily interactions with AI chatbots or virtual assistants?

Consider these practical implications for AI creation and usage:

  • AI for Customer Service: AI might struggle with nuanced customer emotions.
  • AI for Education: Personalized learning could be less effective without true empathy.
  • AI for Creative Writing: Character motivations might lack depth.
  • AI for Therapy: AI cannot genuinely understand a user’s emotional state.

This isn’t to say current AI isn’t useful. However, it highlights a crucial gap. Your expectations for AI should be grounded in these findings. What if AI could truly understand your perspective?

The Surprising Finding

Here’s the twist: Previous research often suggested LLMs did have a form of Theory of Mind. Many evaluations showed success across various social tasks. However, the new study argues these evaluations didn’t test for the actual representations posited by ToM. Instead, they focused on approximating human judgments. The team revealed that while GPT-4o could pass simple tests, it failed at logically identical but structured tasks. This indicates a lack of a “domain-general or consistent ToM,” as mentioned in the release. This finding challenges the common assumption that an LLM’s ability to converse socially implies genuine understanding. It’s surprising because many believed AI was closer to achieving human-like social cognition. The research shows that simply mimicking human behavior isn’t the same as truly understanding the underlying mental processes.

What Happens Next

This research will likely steer future AI creation. Expect to see new evaluation frameworks emerging over the next 12-18 months. These frameworks will focus on probing deeper cognitive abilities, not just surface-level performance. For example, AI researchers might develop more complex scenarios that test an LLM’s ability to reason about nested beliefs (e.g., “John believes that Mary thinks the ball is in the box”). For you, this means a more realistic understanding of AI’s current limitations. If you’re developing AI applications, consider incorporating human oversight for tasks requiring genuine empathy or deep social reasoning. The industry implications are clear: AI might need new architectural approaches to achieve true Theory of Mind. The study’s authors suggest that the social proficiency exhibited by LLMs is not the result of a domain-general or consistent ToM, according to the technical report. This will push the boundaries of AI research significantly.

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice