HiFiTTS-2: Boosting AI Voice with High-Quality Data

A new large-scale dataset promises more realistic and versatile AI-generated speech.

Researchers have introduced HiFiTTS-2, a massive speech dataset derived from LibriVox audiobooks. This dataset aims to enhance the quality of high-bandwidth text-to-speech (TTS) models, making AI voices sound more natural and expressive for various applications.

Sarah Kline

By Sarah Kline

September 23, 2025

4 min read

HiFiTTS-2: Boosting AI Voice with High-Quality Data

Key Facts

  • HiFiTTS-2 is a large-scale speech dataset for high-bandwidth speech synthesis.
  • It contains approximately 36.7k hours of English speech for 22.05 kHz training.
  • The dataset also includes 31.7k hours for 44.1 kHz training.
  • It is derived from LibriVox audiobooks.
  • The dataset comes with detailed utterance and audiobook metadata.

Why You Care

Ever noticed how some AI voices still sound a bit robotic or unnatural? What if your favorite podcast host’s AI clone sounded indistinguishable from the real thing? This is precisely the challenge HiFiTTS-2 aims to solve, bringing us closer to truly lifelike artificial speech. This new dataset is poised to significantly improve how AI generates voices, directly impacting your listening experience.

What Actually Happened

Researchers unveiled HiFiTTS-2, a substantial speech dataset designed for high-bandwidth speech synthesis, according to the announcement. This dataset is built from LibriVox audiobooks. It includes approximately 36.7k hours of English speech for 22.05 kHz training. What’s more, it offers 31.7k hours for 44.1 kHz training, as detailed in the blog post. The team also revealed their data processing pipeline. This pipeline includes steps like bandwidth estimation, segmentation, and text preprocessing. It also incorporates multi-speaker detection. This comprehensive approach ensures high data quality and versatility for various research needs.

Why This Matters to You

This new dataset has practical implications for anyone interacting with AI-generated audio. Imagine listening to an audiobook narrated by an AI voice that perfectly captures human intonation and emotion. This is the future HiFiTTS-2 helps create. It provides the foundational data needed to train more text-to-speech (TTS) models. These models can then produce incredibly realistic speech. The dataset comes with detailed utterance and audiobook metadata. This allows researchers to apply specific data quality filters. This adaptability means the dataset can be tailored for many different use cases, according to the paper.

Here’s how HiFiTTS-2 can enhance AI speech:

  • Higher Fidelity: Enables AI to generate voices with richer, more natural sound quality.
  • Versatile Training: Supports both standard (22.05 kHz) and high-fidelity (44.1 kHz) audio training.
  • Improved Zero-Shot TTS: Facilitates the creation of models that can generate new voices without extensive prior training on that specific voice.
  • Enhanced Expressiveness: Leads to AI voices that better convey emotion and nuance.

“Experimental results demonstrate that our data pipeline and resulting dataset can facilitate the training of high-quality, zero-shot text-to-speech (TTS) models at high bandwidths,” the team revealed. How might more natural AI voices change your daily interactions with system? Think of it as the difference between an old landline call and a crystal-clear HD audio stream. Your experience with voice assistants, audio content, and even virtual characters will become much more engaging.

The Surprising Finding

One interesting aspect of HiFiTTS-2 is its sheer scale and the meticulous processing applied to publicly available LibriVox audiobooks. You might assume that simply compiling hours of audio is enough for AI training. However, the research shows that a detailed data processing pipeline is crucial. This pipeline includes bandwidth estimation and multi-speaker detection. This rigorous approach transforms raw audio into a highly usable resource for AI creation. It challenges the assumption that quantity alone guarantees quality in large datasets. The team specifically highlighted the importance of this processing. This ensures the dataset can train truly high-quality models. This attention to detail is what makes HiFiTTS-2 stand out.

What Happens Next

HiFiTTS-2 was accepted into Interspeech 2025, indicating its significance in the research community. This suggests that further advancements based on this dataset could emerge in late 2025 or early 2026. Researchers will likely use HiFiTTS-2 to develop text-to-speech models. For example, imagine a content creator using AI to voice-over videos in multiple languages with native accents and emotional tones. This dataset provides the foundation for such capabilities. The industry implications are vast, impacting areas from entertainment to accessibility tools. Your future interactions with AI could soon involve voices that are indistinguishable from human speech. The team’s work provides actionable insights for developers. They can now focus on building more realistic and expressive AI voice applications.

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice