AI's Hidden Weakness: Misleading Cues Trip Up MLLMs

New research reveals Multimodal Large Language Models frequently flip correct answers when faced with deceptive information.

Multimodal Large Language Models (MLLMs) are surprisingly vulnerable to misleading information, according to new research. A study found these AI models overturn previously correct answers in over 65% of cases after receiving a single deceptive cue. This 'response uncertainty' highlights a critical challenge for AI reliability.

Mark Ellison

By Mark Ellison

September 16, 2025

4 min read

AI's Hidden Weakness: Misleading Cues Trip Up MLLMs

Key Facts

  • MLLMs overturn correct answers in 65% of cases after a single deceptive cue.
  • Average misleading rates across models exceed 86%.
  • Explicit misleading cues resulted in over 67.19% misleading rate, implicit cues over 80.67%.
  • Fine-tuning with a 2000-sample dataset reduced explicit misleading rates to 6.97%.
  • Fine-tuning boosted consistency by nearly 29.37% on highly deceptive inputs.

Why You Care

Imagine asking an AI a simple question, getting a correct answer, and then asking it again with a subtle, misleading hint. Would you expect the AI to stick to its original, accurate response? A new study reveals that Multimodal Large Language Models (MLLMs) often don’t. This finding could impact your trust in AI systems that combine vision and language. It’s not just about getting facts wrong; it’s about AI changing its mind due to a deceptive nudge. What does this mean for the reliability of the AI tools you use daily?

What Actually Happened

Researchers recently explored a concerning phenomenon in Multimodal Large Language Models (MLLMs). These models excel at tasks like visual question answering, according to the announcement. However, existing studies primarily focused on visual-textual misalignment. The new research, detailed in the paper, investigates MLLMs’ ability to maintain a correct answer when presented with misleading information. The team revealed a significant ‘response uncertainty’ across nine standard datasets. They found that twelve open-source MLLMs overturned a previously correct answer in 65% of cases after just one deceptive cue. This means the AI changed its mind from right to wrong. To quantify this vulnerability, the researchers developed a two-stage evaluation pipeline. First, they elicited each model’s original response. Then, they injected misleading instructions, both explicit (false-answer hints) and implicit (contextual contradictions). They calculated the ‘misleading rate,’ which is the fraction of correct-to-incorrect flips. This systematic approach uncovered a essential flaw in current MLLM design.

Why This Matters to You

This research directly impacts the trustworthiness of AI systems you interact with. If an MLLM can be easily swayed by misleading information, its reliability in essential applications becomes questionable. Think of it as a smart assistant that can be tricked into giving bad advice. For example, imagine using an AI for medical image analysis. If a subtle, misleading text prompt could make it misdiagnose an image, the consequences would be severe. The study highlights that even models are susceptible.

Here’s a breakdown of the observed misleading rates:

Type of Misleading CueAverage Misleading Rate
Overall>86%
Explicit Cues>67.19%
Implicit Cues>80.67%

As mentioned in the release, these rates are alarmingly high. How confident can you be in AI responses if they can be so easily manipulated? One of the authors noted, “We reveal a response uncertainty phenomenon: across nine standard datasets, twelve open-source MLLMs overturn a previously correct answer in 65% of cases after receiving a single deceptive cue.” This indicates a fundamental challenge. Your reliance on AI for factual information could be compromised by these vulnerabilities.

The Surprising Finding

Here’s the twist: While MLLMs are generally considered highly capable, the extent of their ‘response uncertainty’ is quite surprising. Common assumptions suggest that AI models would be against simple deceptions. However, the study finds that average misleading rates exceed 86% across various models. This means that, on average, more than eight out of ten times, an MLLM will flip a correct answer to an incorrect one when given a deceptive hint. This challenges the idea that these models inherently understand and maintain factual consistency. The research specifically highlights that both explicit false-answer hints and implicit contextual contradictions are highly effective at misleading the AI. This suggests a deeper issue than just simple factual errors. It points to a lack of reasoning or confidence in their initial correct assessments.

What Happens Next

The researchers didn’t just identify the problem; they also explored solutions. To reduce this ‘misleading rate,’ they fine-tuned all open-source MLLMs. They used a compact 2000-sample mixed-instruction dataset, according to the announcement. This intervention significantly reduced misleading rates. For explicit cues, the rate dropped to 6.97%. For implicit cues, it fell to 32.77%. This boosted consistency by nearly 29.37% on highly deceptive inputs. It also slightly improved accuracy on standard benchmarks, the study finds. This suggests that targeted training can make MLLMs more resilient. In the coming months, expect to see more research focusing on making MLLMs . For example, future AI assistants might incorporate similar fine-tuning techniques to prevent easy manipulation. For you, this means potentially more reliable AI tools in the near future. Industry implications include a push for more AI safety and reliability standards. The team revealed that their code is available, encouraging further research and creation in this essential area.

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice