Unpacking AI's Brain: How LLMs Handle Language Tasks

New research explores the distinct neural pathways large language models use for formal grammar versus complex reasoning.

A recent study investigates whether large language models (LLMs) process formal linguistic tasks (like grammar) and functional linguistic tasks (like reasoning) using separate internal mechanisms. The findings suggest that while there's little overlap between these task types, the internal 'circuits' for formal tasks also show surprising independence, challenging previous assumptions about how LLMs work.

Katie Rowan

By Katie Rowan

August 31, 2025

4 min read

Unpacking AI's Brain: How LLMs Handle Language Tasks

Key Facts

  • Researchers investigated whether LLMs use distinct internal mechanisms for formal and functional linguistic tasks.
  • LLMs excel at formal tasks (e.g., fluent text) but struggle with functional tasks (e.g., reasoning, fact retrieval).
  • The study compared 5 LLMs across 10 distinct tasks, analyzing their internal 'circuits'.
  • Findings show little overlap between formal and functional task circuits.
  • Surprisingly, there was also little overlap between different formal linguistic tasks, suggesting no single unified formal network.

Why You Care

Have you ever wondered if your favorite AI chatbot truly ‘understands’ what you’re asking, or if it’s just really good at mimicking human speech? A new paper, as mentioned in the release, dives deep into the internal workings of large language models (LLMs).

This research, published by Michael Hanna and his colleagues, looks at how these AI systems handle different types of language tasks. Understanding this could change how we build future AI. It directly impacts the reliability and intelligence of the AI tools you use every day.

What Actually Happened

Researchers Michael Hanna, Yonatan Belinkov, and Sandro Pezzelle investigated whether LLMs use distinct internal mechanisms for formal and functional linguistic tasks. According to the announcement, LLMs generally excel at formal tasks, such as generating fluent and grammatically correct text. However, the study finds they often struggle more with functional tasks, which involve reasoning and consistent fact retrieval.

To explore this, the team compared five different LLMs across ten distinct tasks. They looked for ‘circuits,’ or minimal computational subgraphs, responsible for these tasks. These circuits are essentially the specific parts of the model’s internal network that activate for certain operations. The paper states that their goal was to see if these models exhibit distinct localization of formal and functional linguistic mechanisms.

Why This Matters to You

This distinction between formal and functional mechanisms has significant implications for how you interact with AI. If an AI struggles with reasoning, your conversations might feel less intelligent or reliable. Imagine you’re using an AI assistant to plan a complex trip. You expect it to understand your preferences, reason about travel times, and retrieve accurate information, not just generate a grammatically but nonsensical itinerary.

This research helps us understand the limitations and strengths of current AI. It suggests areas where AI creation needs to focus to improve overall performance. What if future AI could reason as flawlessly as it writes? How would that change your daily life?

Consider these key differences in LLM capabilities:

Task TypeExamples of Capability
Formal LinguisticProducing fluent, grammatical sentences; correcting syntax; generating creative text.
Functional LinguisticAnswering complex questions; maintaining factual consistency; logical deduction; problem-solving.

As detailed in the blog post, the ability of one circuit to solve another’s task, known as cross-task faithfulness, showed a clear separation. This indicates that shared mechanisms between formal tasks might exist, even if a single unified formal network remains elusive. According to the research, “Although large language models (LLMs) are increasingly capable, these capabilities are unevenly distributed: they excel at formal linguistic tasks, such as producing fluent, grammatical text, but struggle more with functional linguistic tasks like reasoning and consistent fact retrieval.”

The Surprising Finding

Here’s the twist: the researchers expected to find a clear separation between formal and functional tasks, perhaps with formal tasks sharing a common internal network. However, the study revealed something more nuanced. The team revealed that while there was indeed little overlap between the circuits for formal and functional tasks, there was also surprisingly little overlap between different formal linguistic tasks themselves.

This challenges a common assumption, inspired by neuroscience, that a single formal linguistic network might exist within LLMs. The study finds that a “single formal linguistic network, unified and distinct from functional task circuits, remains elusive.” This means that even within the realm of grammar and fluency, different formal aspects might be handled by largely independent internal components. It’s like finding out that your car has separate tiny engines for accelerating, braking, and steering, instead of one central system for movement.

The research compared 5 LLMs across 10 distinct tasks.

What Happens Next

This understanding of internal AI mechanisms will guide future AI creation. Developers might focus on building more integrated or specialized circuits for specific functional tasks. For example, future models could incorporate dedicated ‘reasoning engines’ that are distinct from their ‘grammar engines.’

Actionable advice for you: as an AI user, be aware of these distinctions. Don’t assume an AI’s grammatical fluency equates to deep understanding or factual accuracy. Always verify essential information provided by LLMs. The documentation indicates that code for this research is available, allowing other researchers to build upon these findings.

In the coming months, perhaps by late 2025 or early 2026, we could see new LLM architectures emerge. These might explicitly design for separate or more functional mechanisms. This will lead to AI that is not only eloquent but also genuinely intelligent in its reasoning. The industry implications are significant, pushing us towards more reliable and AI assistants.

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice