AI Models Show Brain-Like Patterns After Instruction-Tuning

New research reveals how instruction-tuned multimodal LLMs align with human brain activity.

A recent study found that instruction-tuned multimodal large language models (IT-MLLMs) exhibit stronger alignment with human brain activity compared to other AI models. This alignment suggests IT-MLLMs process information in ways similar to how our brains handle specific tasks. The findings could lead to more brain-inspired AI.

Mark Ellison

By Mark Ellison

February 10, 2026

4 min read

AI Models Show Brain-Like Patterns After Instruction-Tuning

Key Facts

  • Instruction-tuned multimodal LLMs (IT-MLLMs) show stronger alignment with human brain activity.
  • The study used fMRI responses recorded during naturalistic movie watching to evaluate brain alignment.
  • IT-MLLMs outperformed in-context learning, non-instruction-tuned, and unimodal models in brain alignment.
  • ICL models showed strong semantic organization (r=0.78), while IT models showed weak coupling to instruction-text semantics (r=0.14).
  • The research suggests IT-MLLMs organize representations around functional task demands, not just surface semantics.

Why You Care

Ever wondered if artificial intelligence could really think like you do? What if AI models could process information in ways that mirror your own brain’s activity? New research suggests this future might be closer than we think. This creation could reshape how we design AI, making it more intuitive and for your daily tasks.

What Actually Happened

A team of researchers, including Subba Reddy Oota and Manish Gupta, recently published a study on arXiv. The paper, titled “Task-Conditioned Probing Reveals Brain-Alignment Patterns in Instruction-Tuned Multimodal LLMs,” explores how instruction-tuned multimodal large language models (IT-MLLMs) align with human brain activity. These IT-MLLMs are AI models that can understand and process different types of data, like video and audio, based on specific instructions. The research focused on predicting fMRI responses—functional magnetic resonance imaging, which measures brain activity—recorded during naturalistic movie watching, as mentioned in the release. They used instruction-specific embeddings from six video and two audio IT-MLLMs across 13 video task instructions, according to the announcement.

Why This Matters to You

This research indicates that IT-MLLMs don’t just understand instructions; they organize their internal representations in a way that aligns with how your brain processes tasks. Imagine an AI assistant that truly grasps the nuances of your requests. For example, if you ask an IT-MLLM to summarize a video, it might process that video’s content similarly to how your brain would, focusing on key elements. This could lead to AI tools that are far more intuitive and responsive to your needs. The study finds that IT-MLLMs significantly outperform other models in brain alignment.

Performance Comparison in Brain Alignment:

  • Instruction-Tuned Video MLLMs: Outperformed in-context learning (ICL) multimodal models by approximately 9%.
  • Instruction-Tuned Video MLLMs: Outperformed non-instruction-tuned multimodal models by approximately 15%.
  • Instruction-Tuned Video MLLMs: Outperformed unimodal baselines by approximately 20%.

How might this improved brain alignment change your interaction with AI in the coming years? The team revealed that “instruction-tuned video MLLMs significantly outperform in-context learning (ICL) multimodal models (~9%), non-instruction-tuned multimodal models (~15%), and unimodal baselines (~20%).” This means AI could become much better at understanding complex, multi-step instructions, just like your brain does.

The Surprising Finding

Here’s an interesting twist: the study found a surprising difference in how various models handle semantics. While in-context learning (ICL) models showed strong semantic organization, with a correlation of r=0.78, instruction-tuned (IT) models displayed weak coupling to instruction-text semantics, with a correlation of r=0.14. This might seem counterintuitive at first. You might assume that models better aligned with the brain would also have a stronger semantic link to their instructions. However, the paper states this weak coupling in IT models is consistent with “task-conditioned subspaces associated with higher brain alignment.” It suggests that IT-MLLMs aren’t just mimicking surface-level meaning. Instead, they are developing deeper, task-specific internal structures that are more akin to how the brain organizes information for specific functions. This challenges the common assumption that strong semantic literalism is always best for brain-like processing.

What Happens Next

This research opens up exciting possibilities for the future of AI. The code for this study has been made publicly available, which means other researchers can build upon these findings. We could see advancements in brain-inspired AI within the next 12-18 months. Imagine AI systems for medical diagnostics that interpret images with a human-like understanding, or virtual assistants that anticipate your needs based on subtle cues. The industry implications are vast, potentially leading to more AI in areas like robotics, natural language processing, and computer vision. For you, this means future AI tools could be more intuitive and effective, requiring less explicit instruction. The team revealed these findings “open new avenues for mapping joint information processing in both systems.”

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice