AI Voices Get Emotional Control with New LibriTTS-VI Corpus

Researchers tackle 'impression leakage' in text-to-speech, offering fine-tuned emotional expression.

A new public dataset, LibriTTS-VI, and novel methods are improving how AI voices convey emotions. This advancement helps text-to-speech systems produce more natural and controllable voice impressions, addressing key challenges like 'impression leakage.'

Mark Ellison

By Mark Ellison

September 23, 2025

4 min read

AI Voices Get Emotional Control with New LibriTTS-VI Corpus

Key Facts

  • Researchers introduced LibriTTS-VI, a new public corpus for voice impression control.
  • New methods address 'impression leakage' in text-to-speech synthesis.
  • The best method reduced objective mean squared error of voice impression vectors from 0.61 to 0.41.
  • Subjective mean squared error also improved from 1.15 to 0.92.
  • A novel reference-free model generates speaker embeddings directly from target impressions.

Why You Care

Ever wished your AI assistant could sound genuinely happy, or perhaps a bit more serious, without sounding robotic? Imagine the difference that would make in your daily interactions. This is precisely what new research is bringing closer to reality. A team of researchers has unveiled significant advancements in controlling the emotional nuances of AI-generated voices. This means your future interactions with AI could feel much more natural and expressive.

What Actually Happened

Researchers have introduced a new public corpus called LibriTTS-VI, alongside novel methods for efficient voice impression control. This creation addresses two main challenges in creating more controllable text-to-speech (TTS) systems, according to the announcement. The first challenge is “impression leakage,” where a synthesized voice unintentionally picks up emotional characteristics from the reference audio. The second is the previous lack of a publicly available, annotated dataset for voice impressions.

To combat impression leakage, the team proposed two methods. One involves a training strategy that uses separate utterances for speaker identity and for the target impression, as detailed in the blog post. The other is a “novel reference-free model” which generates a speaker embedding — a unique digital fingerprint of a voice — directly from the desired impression. This approach improves robustness and offers the convenience of generating voices without needing a reference audio sample.

Why This Matters to You

This research significantly improves the ability of AI to generate voices with specific emotional tones. Think about how this could change audiobooks, virtual assistants, or even your podcast productions. You could specify an AI voice to sound ‘brighter’ or ‘calmer,’ tailoring it to your exact needs.

For example, imagine you are producing a podcast. Instead of a monotone AI voice, you could have a voice actor AI express enthusiasm for a product review or a somber tone for a news segment. This level of control opens up new creative possibilities for content creators.

Key Improvements in Voice Impression Control:

  • Reduced Impression Leakage: AI voices are less likely to pick up unwanted emotional cues from source audio.
  • Enhanced Controllability: Users can specify desired emotional impressions like ‘brighter’ or ‘calmer.’
  • Public Dataset Availability: LibriTTS-VI provides a standardized resource for further research and creation.
  • Reference-Free Generation: New models can create emotional voices without needing a specific audio example.

How might this enhanced emotional control change the way you interact with AI in your professional or personal life? The study finds that objective evaluations demonstrated significant betterment in controllability. “Objective and subjective evaluations demonstrate a significant betterment in controllability,” the paper states, highlighting the effectiveness of their new methods.

The Surprising Finding

Perhaps the most surprising aspect of this research is the substantial objective betterment in voice impression control. The team revealed that their best method significantly reduced the mean squared error (MSE) of 11-dimensional voice impression vectors. This error dropped from 0.61 to 0.41 objectively, and from 1.15 to 0.92 subjectively, all while maintaining high fidelity. This means the AI’s ability to accurately reproduce a desired emotional impression improved dramatically, far beyond what might be expected from initial attempts in such a nascent field.

This finding challenges the common assumption that fine-grained emotional control in AI voices is a distant goal. The researchers achieved concrete, measurable progress. It shows that addressing core technical issues like impression leakage can yield results. It’s not just about making voices sound human; it’s about making them sound specifically human, with precise emotional intent.

What Happens Next

This research, submitted to ICASSP 2026, suggests that we can expect further developments in this area over the next few years. The availability of LibriTTS-VI, described as “the first public voice impression dataset released with clear annotation standards,” will likely accelerate research. This public resource, built upon the LibriTTS-R corpus, will enable more researchers to build upon these findings.

For example, developers could soon integrate these voice impression controls into their text-to-speech APIs, offering creators flexibility. If you are a content creator, you might soon have access to tools that allow you to dictate not just what an AI voice says, but how it says it, with emotional precision. The industry implications are vast, from more engaging virtual assistants to more immersive audio experiences. The team hopes to foster reproducible research, as mentioned in the release, pushing the boundaries of what AI voices can achieve.

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice