New Benchmark Tackles LLM Hallucinations

DeepMind's FACTS Grounding dataset evaluates how accurately large language models stick to source material.

DeepMind has launched FACTS Grounding, a new benchmark and leaderboard designed to measure the factual accuracy of large language models (LLMs). This initiative aims to reduce 'hallucinations' by evaluating how well LLMs ground their responses in provided source documents. The dataset includes 1,719 examples across diverse domains.

Katie Rowan

By Katie Rowan

December 4, 2025

4 min read

New Benchmark Tackles LLM Hallucinations

Key Facts

  • DeepMind introduced FACTS Grounding, a new benchmark for evaluating LLM factuality.
  • The benchmark aims to reduce LLM 'hallucinations' by ensuring responses are grounded in source material.
  • The FACTS Grounding dataset contains 1,719 examples requiring long-form responses.
  • LLM responses are automatically evaluated by three frontier LLM judges: Gemini 1.5 Pro, GPT-4o, and Claude 3.5 Sonnet.
  • The dataset includes documents up to 32,000 tokens (approx. 20,000 words) across diverse domains.

Why You Care

Ever asked an AI a question, only to get a confident-sounding answer that was completely wrong? It’s frustrating, right? This problem, where large language models (LLMs) invent information, is called ‘hallucination’. Now, a new benchmark aims to fix it. DeepMind has introduced FACTS Grounding, a essential tool for evaluating and improving the factual accuracy of these AI systems. Why should you care? Because your trust in AI, and its usefulness in your daily life, depends on its ability to tell you the truth.

What Actually Happened

DeepMind has launched FACTS Grounding, a new benchmark specifically designed to assess the factuality of large language models, according to the announcement. This initiative directly addresses the issue of LLMs generating false information, especially with complex inputs. The team revealed that this tendency can erode trust and limit AI’s real-world applications. To combat this, the FACTS Grounding dataset comprises 1,719 carefully crafted examples. Each example requires LLMs to produce long-form responses strictly based on a provided context document. The benchmark also includes an online leaderboard to track industry-wide progress on factuality and grounding, as mentioned in the release. This provides a public way to see which models are performing best.

Why This Matters to You

This new benchmark directly impacts the reliability of the AI tools you use every day. Imagine using an AI for research or even just answering a quick question. You need to know you can trust its output. The FACTS Grounding dataset helps ensure that LLMs stick to the facts. For example, if you ask an AI to summarize a financial report, this benchmark helps guarantee the summary only contains information from that report, not invented details.

Key Features of FACTS Grounding Dataset:

  • Total Examples: 1,719
  • Public Set: 860 examples
  • Private Set: 859 examples
  • Max Token Length: 32,000 tokens (approximately 20,000 words)
  • Domains Covered: Finance, system, retail, medicine, law

“We hope our benchmark will spur industry-wide progress on factuality and grounding,” the team stated, highlighting their ambition for widespread adoption. This means better, more trustworthy AI for everyone. Do you ever double-check AI-generated information because you’re unsure of its accuracy? This benchmark aims to reduce that need, making your interactions with AI smoother and more dependable.

The Surprising Finding

Interestingly, the benchmark design itself reveals a subtle but important twist in how factuality is being approached. Instead of relying solely on human judgment, FACTS Grounding evaluates model responses automatically. It uses three frontier LLM judges: Gemini 1.5 Pro, GPT-4o, and Claude 3.5 Sonnet. This collective judgment by leading LLMs is a clever way to mitigate potential bias. The technical report explains that this approach avoids any single judge giving higher scores to its own model family’s responses. This method challenges the assumption that only human experts can reliably assess factual accuracy. The team comprehensively evaluated these automatic judge models against a held-out test set. This confirmed their agreement with human raters, providing confidence in this evaluation strategy.

What Happens Next

The introduction of FACTS Grounding and its accompanying leaderboard sets a clear direction for AI creation. We can expect to see LLM developers actively competing to improve their models’ scores on this benchmark in the coming months. For example, companies might release updated versions of their models, perhaps by early to mid-next year, specifically touting their improved grounding capabilities. This will likely lead to more reliable AI assistants and content generation tools. For you, this means future AI interactions will be less prone to frustrating ‘hallucinations’. The industry implications are significant, pushing all LLM providers towards greater transparency and accuracy. This will ultimately enhance the trustworthiness of AI across various applications. The team emphasized the importance of this work, stating, “To succeed on a given example, an LLM must synthesize the complex information in the document and generate a long-form response that is both a comprehensive answer to the user request and fully attributable to that document.”

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice