AI Solves Historical Text Puzzles with New Confidence-Based System

MHEL-LLaMo leverages LLMs and SLMs to accurately link entities in old documents.

Researchers have developed MHEL-LLaMo, an unsupervised AI system for historical entity linking. This system combines large and small language models to efficiently identify and link historical entities across multiple languages, overcoming challenges like linguistic variation and noisy data without requiring extensive training.

Sarah Kline

By Sarah Kline

January 20, 2026

4 min read

AI Solves Historical Text Puzzles with New Confidence-Based System

Key Facts

  • MHEL-LLaMo is an unsupervised approach for multilingual historical entity linking.
  • It combines a Small Language Model (SLM) and a Large Language Model (LLM).
  • The system uses SLM confidence scores to decide when to apply the LLM.
  • MHEL-LLaMo works on texts from the 19th and 20th centuries across six European languages.
  • It outperforms state-of-the-art models without requiring fine-tuning.

Why You Care

Ever struggle to understand old documents or historical texts? Imagine an AI that can accurately connect names, places, and events in those challenging records. How much easier would your historical research become?

A new system called MHEL-LLaMo is changing how we interact with historical data. It offers an unsupervised approach to multilingual historical entity linking. This means it can identify and link entities in old texts without needing human-labeled examples. For content creators, podcasters, or anyone dealing with historical archives, this could dramatically simplify your work.

What Actually Happened

Researchers Cristian Santini, Marieke Van Erp, and Mehwish Alam have introduced MHEL-LLaMo. This system tackles the tough problem of historical entity linking (EL), as detailed in the blog post. Historical texts often contain linguistic variations, noise, and evolving semantic conventions. These factors make traditional entity linking difficult.

MHEL-LLaMo is an unsupervised ensemble approach. It combines a Small Language Model (SLM) with a Large Language Model (LLM). The system uses a multilingual bi-encoder (BELA) to find potential candidates. Then, an instruction-tuned LLM handles NIL prediction and candidate selection. This process uses prompt chaining, according to the announcement. The key creation lies in using the SLM’s confidence scores. This helps discriminate between easy and hard samples. The LLM is only applied to the more challenging cases.

Why This Matters to You

This new approach offers significant practical implications. It reduces computational costs, as the team revealed. By only using the LLM for complex tasks, it prevents “hallucinations” on straightforward cases. This makes the system more efficient and reliable. Think of it as having a junior researcher handle routine tasks, while a senior expert steps in for the tough questions.

MHEL-LLaMo’s Key Advantages:

  • Unsupervised Learning: No need for extensive, human-labeled training data.
  • Multilingual Support: Works across six European languages.
  • Cost-Efficiency: Reduces computational load by selectively using LLMs.
  • Improved Accuracy: Outperforms models without fine-tuning.

Imagine you are creating a podcast about 19th-century European history. You encounter a diary entry mentioning a ‘Monsieur Dubois.’ MHEL-LLaMo could help you quickly identify if this refers to a famous general, a local baker, or another historical figure. How much time could this save your research efforts?

As the paper states, MHEL-LLaMo “outperforms models without requiring fine-tuning, offering a approach for low-resource historical EL.” This means you get better results with less upfront effort.

The Surprising Finding

The most surprising aspect of MHEL-LLaMo is its confidence-based strategy. The system uses a Small Language Model’s confidence scores to decide when to engage the more resource-intensive Large Language Model. This is counterintuitive because you might expect to always use the most AI available. However, the technical report explains that this method actually improves efficiency and accuracy.

This strategy prevents the LLM from generating incorrect information, known as “hallucinations,” on simpler tasks. By relying on the SLM for easy cases, the system avoids unnecessary processing. It also ensures higher quality output. This challenges the common assumption that more models are always better for every part of a task. Instead, smart allocation of AI resources proves more effective.

What Happens Next

The implementation of MHEL-LLaMo is already available on GitHub, as mentioned in the release. This means developers and researchers can start experimenting with it immediately. We can expect to see early integrations and tests within the next 3-6 months. Over the next year, this system could be adopted by digital humanities projects and archival institutions.

For example, a historical society might use MHEL-LLaMo to process digitized newspaper archives. This would allow them to automatically link historical figures and events. This could create rich, interconnected databases for public access. If you work with historical documents, consider exploring the GitHub repository. Staying informed about such advancements can give your projects a significant edge. This creation offers a approach for historical entity linking, according to the authors. It promises to make historical data more accessible and understandable for everyone.

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice