New AI Fights Deepfakes with Explainable Detection

Researchers propose MLLM-powered framework for transparent fake image identification.

A new study introduces an explainable framework for detecting AI-generated fake images using Multi-modal Large Language Models (MLLMs). This approach aims for both strong detection accuracy and transparency, moving beyond 'black box' methods. It offers a more reliable way to identify deepfakes.

Sarah Kline

By Sarah Kline

November 11, 2025

4 min read

New AI Fights Deepfakes with Explainable Detection

Key Facts

  • AI-generated image progress raises public security concerns.
  • Fake image detection needs to be transparent, not a 'black box'.
  • Multi-modal Large Language Models (MLLMs) offer new reasoning-based detection opportunities.
  • Researchers evaluated MLLMs against traditional methods and human evaluators.
  • A framework with six distinct prompts was designed for robust, explainable detection.

Why You Care

Ever worry if the images you see online are real or AI-generated fakes? With rapid advancements in image creation system, it’s getting harder to tell. This creation is crucial for your digital safety and trust. A new research paper tackles this problem head-on. It offers a fresh perspective on identifying fake images. This approach promises not just detection, but also explanations. How might this change your daily online experience?

What Actually Happened

Progress in AI image generation has raised significant public security concerns. This is according to the announcement from a team of researchers. They argue that fake image detection should not operate as a “black box.” Instead, an ideal approach needs both strong generalization and transparency. Recent advancements in Multi-modal Large Language Models (MLLMs) — AI models that can process and understand multiple types of data like text and images — offer new opportunities. These opportunities are for reasoning-based AI-generated image detection. The research team evaluated MLLMs’ capabilities. They compared them to traditional detection methods and human evaluators. This comparison highlighted both strengths and limitations of current MLLM approaches. What’s more, the team designed six distinct prompts. They then proposed a structure integrating these prompts. This structure aims to develop a more , explainable, and reasoning-driven detection system. The code for this system is available, as mentioned in the release.

Why This Matters to You

Imagine you’re scrolling through social media. You see a shocking image. Is it real news or a cleverly crafted deepfake? This new structure could help you make that distinction. It goes beyond simply labeling an image as fake. It explains why it’s fake. This transparency builds trust in the detection process. The study finds that MLLMs offer unique advantages over older methods. “We argue that fake image detection should not operate as a ‘black box’. Instead, an ideal approach must ensure both strong generalization and transparency,” the paper states. This means you get a clear explanation, not just a yes or no answer. This is vital for essential decision-making. How much more confident would you be sharing information if you knew it was by an explainable AI?

Here are some key benefits for you:

  • Increased Trust: Understand why an image is flagged as fake.
  • Better Informed Decisions: Make more confident judgments about online content.
  • Enhanced Security: Reduce the risk of falling for deepfakes.
  • Improved Digital Literacy: Learn to identify common deepfake tells through explanations.

For example, if an MLLM flags an image, it might explain that “the lighting on the subject’s face is inconsistent with the background light source.” This level of detail empowers you. It moves beyond simple algorithmic verdicts. This approach makes fake image detection more accessible and understandable for everyone.

The Surprising Finding

Here’s an interesting twist: the research shows that while MLLMs offer new opportunities, they also have limitations compared to human evaluators. The team revealed that MLLMs aren’t out of the box. They still require careful prompting and structure design. This challenges the assumption that AI can instantly solve complex problems. The study highlights the need for a structured approach. MLLMs, when properly prompted, can provide reasoning-based detection. This is a crucial step beyond simple classification. It means that the way we ask AI to detect fakes matters significantly. The researchers’ structure with six distinct prompts is key to unlocking this explainability. It’s not just about raw AI power. It’s about intelligent design and interaction. This finding suggests that human ingenuity in crafting AI interactions remains vital. It is essential even with models.

What Happens Next

This research, accepted to ACM MM 2025, indicates future developments. We can expect more refined explainable fake image detection systems to emerge. The team’s work provides a foundation. For example, imagine a browser extension that not only flags suspicious images but also provides a concise explanation. This could be available within the next 12-18 months. Industry implications are significant. Content platforms and social media companies could integrate these explainable MLLM frameworks. This would enhance their content moderation efforts. For you, this means a safer and more transparent online environment. Keep an eye out for tools that incorporate these reasoning-driven approaches. Your ability to discern real from fake online content will only improve. The documentation indicates that the code is already available. This suggests that further creation and adoption could happen relatively quickly.

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice