CAPO Boosts Multilingual LLM Alignment by 16%

New method improves large language models' understanding of human preferences across languages.

A new technique called Confidence-Aware Preference Optimization (CAPO) significantly improves how large language models (LLMs) align with human preferences in multilingual settings. It addresses limitations of existing methods like DPO by dynamically scaling learning based on confidence in preference data. This leads to more robust and accurate LLMs for global use.

Mark Ellison

By Mark Ellison

November 23, 2025

4 min read

CAPO Boosts Multilingual LLM Alignment by 16%

Key Facts

  • CAPO stands for Confidence-Aware Preference Optimization.
  • CAPO addresses the limitations of DPO (Direct Preference Optimization) in multilingual settings.
  • It uses a dynamic loss scaling mechanism based on relative reward.
  • CAPO outperforms existing baselines by at least 16% in reward accuracy.
  • The research was accepted at IJCNLP-AACL 2025 Findings.

Why You Care

Ever wonder why your favorite AI chatbot sometimes struggles with languages other than English? Or why its responses feel a bit ‘off’ when you switch to French or Spanish? This isn’t just a minor inconvenience; it points to a significant challenge in making AI truly global. A new creation could change your experience dramatically.

Researchers have introduced CAPO, or Confidence-Aware Preference Optimization. This creation aims to make large language models (LLMs) much better at understanding and responding to human preferences across many languages. This means more accurate, more helpful, and more culturally aware AI interactions for you, no matter what language you speak.

What Actually Happened

Scientists have developed a new technique called Confidence-Aware Preference Optimization (CAPO). This method is designed to enhance how large language models (LLMs) learn from human feedback, especially in diverse language environments, according to the announcement. LLMs are AI systems that generate text, translate languages, and answer questions.

Traditionally, a method called Direct Preference Optimization (DPO) has been effective for aligning LLMs with human preferences, the research shows. However, DPO often falls short when dealing with multiple languages. CAPO offers a “simple yet effective alternative” to DPO, as detailed in the blog post. It replaces DPO’s fixed approach to preference pairs with a dynamic loss scaling mechanism. This mechanism uses a ‘relative reward’ to adjust how much the model learns from each piece of feedback. This makes CAPO more resilient to noisy or less clear preference data, which is common in multilingual text.

Why This Matters to You

Imagine you’re a content creator trying to reach a global audience. You rely on AI tools to help draft content in various languages. If the AI doesn’t truly grasp the nuances of each language, your message can get lost or misinterpreted. CAPO directly tackles this problem, offering a path to more reliable multilingual AI.

CAPO’s Key Advantages:

  • Enhanced Robustness: It handles noisy preference data better, especially in diverse language sets.
  • Improved Alignment: CAPO widens the gap between preferred and dispreferred responses. This means the AI is clearer about what humans like and dislike.
  • Superior Accuracy: The method outperforms existing baselines in reward accuracy.

Think of it as teaching an AI to be a better listener, not just in English, but in every language. “By modulating the learning signal according to the confidence in each preference pair, CAPO enhances robustness to noisy or low-margin comparisons, typically encountered in multilingual text,” the paper states. This leads to more accurate and culturally sensitive AI outputs. How might more accurate multilingual AI change the way you interact with system daily?

The Surprising Finding

Here’s the twist: while existing methods like DPO work well for English, their effectiveness often drops significantly in multilingual scenarios. This might seem counterintuitive since LLMs are often touted as universal language tools. However, the study finds that CAPO dramatically improves performance where others struggle.

Specifically, CAPO “outperforms existing preference optimization baselines by at least 16% in reward accuracy,” the team revealed. This isn’t a small betterment; it’s a substantial leap in how well LLMs can learn human preferences across different languages. It challenges the assumption that a one-size-fits-all approach to preference optimization works for all linguistic contexts. The dynamic loss scaling mechanism is key here, allowing the model to adapt its learning based on the reliability of the feedback it receives.

What Happens Next

The creation of CAPO marks a significant step for large language models. We can expect to see this method integrated into commercial LLMs over the next 12 to 18 months, according to industry experts. This could mean more translation services and AI assistants that understand cultural subtleties.

For example, imagine an AI customer service bot that can not only speak multiple languages but also understand the specific cultural context of a customer’s query. This would lead to much more satisfying and effective interactions. Developers should consider exploring CAPO for their multilingual AI projects. The company reports that CAPO “improves alignment by widening the gap between preferred and dispreferred responses across languages.” This suggests a future with more nuanced and globally aware AI. Your next AI interaction could be much more intelligent, no matter the language.

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice