New Method Shields LLMs from Forgetting Old Skills

Source-Shielded Updates (SSU) helps large language models learn new languages without losing their original abilities.

Researchers have introduced Source-Shielded Updates (SSU), a technique that allows large language models (LLMs) to adapt to new languages using unlabeled data without 'catastrophic forgetting.' This method selectively protects critical parameters, significantly reducing performance degradation on original tasks while improving target language performance.

Mark Ellison

By Mark Ellison

December 18, 2025

4 min read

New Method Shields LLMs from Forgetting Old Skills

Key Facts

  • Source-Shielded Updates (SSU) mitigates catastrophic forgetting in LLMs during target language adaptation.
  • SSU uses a selective parameter update strategy, freezing critical source knowledge parameters.
  • The method relies only on unlabeled target language data, addressing low-resource constraints.
  • SSU reduced performance degradation on source tasks to 3.4% (7B) and 2.8% (13B) on average, compared to over 20% for full fine-tuning.
  • SSU achieves target-language performance competitive with, and often better than, full fine-tuning.

Why You Care

Ever feel like learning something new makes you forget something old? Large language models (LLMs) face a similar challenge when adapting to new languages. How can we teach an AI a new skill without it losing its core knowledge?

This is a major hurdle for making AI accessible globally. A new approach called Source-Shielded Updates (SSU) directly addresses this. It helps LLMs learn new languages efficiently, keeping their original abilities intact. This means more diverse and capable AI tools for everyone, including you.

What Actually Happened

A team of researchers, including Atsuki Yamaguchi and Terufumi Morishita, has developed a novel strategy to combat “catastrophic forgetting” in large language models, according to the announcement. This issue arises when LLMs are adapted to new target languages, often leading to a significant loss of performance on their original, source language tasks. The new method, Source-Shielded Updates (SSU), is designed to preserve the model’s existing knowledge.

As detailed in the blog post, SSU works by selectively updating parameters. It identifies which parts of the model are crucial for maintaining its initial capabilities. Then, it ‘freezes’ these important parameters before the model begins adapting to a new language. This process uses only unlabeled target language data, which is a key advantage given the high cost of labeled data. The research shows that SSU effectively mitigates this forgetting problem.

Why This Matters to You

Imagine you’re using an AI assistant that understands English perfectly. Now, imagine you want it to also understand and respond in Japanese. Without SSU, adapting that AI to Japanese might make it forget how to speak English well. This new method changes that. It means your AI tools can become multilingual without sacrificing their original proficiency.

This is especially important for expanding the linguistic diversity of instruct LLMs, as mentioned in the release. It allows these models to be used by more people around the world. The study finds that SSU achieves target-language performance highly competitive with full fine-tuning, often outperforming it. This means better-performing models for you, even in low-resource language settings.

Key Benefits of Source-Shielded Updates (SSU):

  • Preserves Source Knowledge: Significantly reduces performance loss on original tasks.
  • Low-Resource Adaptation: Uses only unlabeled target language data, cutting costs.
  • Competitive Performance: Achieves strong results in target languages.
  • Global Accessibility: Enables LLMs to support a wider range of languages.

How much better could your multilingual AI interactions be with this system? The team revealed that SSU successfully mitigates catastrophic forgetting. “It reduces performance degradation on monolingual source tasks to just 3.4% (7B) and 2.8% (13B) on average,” a stark contrast to traditional methods.

The Surprising Finding

Here’s the twist: traditional thinking suggests that to make an LLM learn a new language, you need to retrain a significant portion of it. This often leads to the model ‘forgetting’ some of its original language skills. However, the research shows that SSU dramatically reduces this problem.

The study finds that full fine-tuning resulted in a performance degradation of 20.3% for 7B models and 22.3% for 13B models on source tasks. In stark contrast, SSU limited this degradation to just 3.4% for 7B models and 2.8% for 13B models on average. This is surprising because it demonstrates that you don’t need to sacrifice core abilities to gain new ones. It challenges the assumption that extensive retraining is always necessary. By being selective, the model retains its foundational knowledge much more effectively.

What Happens Next

This research, submitted on December 4, 2025, points to a future where large language models are truly multilingual and globally accessible. We can expect to see these techniques integrated into commercial LLMs within the next 12 to 18 months. This could mean more and versatile AI assistants available to you.

For example, imagine a customer service chatbot. It could seamlessly switch between English, Spanish, and Japanese, without ever stumbling on its core functions. This would enhance user experience significantly. The industry implications are clear: a path to more efficient and less costly multilingual AI creation. The paper states that SSU outperforms full fine-tuning on all benchmarks for 7B models and the majority for 13B models. This suggests a new standard for language adaptation. Developers will likely adopt these selective update strategies. This will help them build more capable and less ‘forgetful’ AI systems for your use.

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice