FlexiVoice: AI Achieves Unprecedented Voice Control

New text-to-speech system offers flexible style and zero-shot voice cloning via natural language.

Researchers have unveiled FlexiVoice, an advanced text-to-speech (TTS) system. It allows users to control speaking style and voice timbre using simple natural language instructions and a short audio reference. This development represents a significant step forward in realistic and customizable AI-generated speech.

Katie Rowan

By Katie Rowan

January 9, 2026

4 min read

FlexiVoice: AI Achieves Unprecedented Voice Control

Key Facts

  • FlexiVoice is a text-to-speech (TTS) system offering flexible style control and zero-shot voice cloning.
  • It uses a Large Language Model (LLM) core for processing text and instructions.
  • Speaking style is controlled by natural language instructions.
  • Voice timbre is provided by a speech reference in a zero-shot manner.
  • A Progressive Post-Training (PPT) scheme, including DPO and GRPO, enables its control capabilities.

Why You Care

Ever wished you could make an AI voice sound exactly how you want, just by telling it? What if you could clone any voice with a tiny audio clip and then dictate its emotional tone? This new creation in text-to-speech (TTS) system is making that a reality. It promises to transform how you interact with AI voices, offering control and realism.

What Actually Happened

A team of researchers has introduced FlexiVoice, a novel text-to-speech synthesis system, according to the announcement. This system stands out for its ability to offer flexible style control alongside zero-shot voice cloning. Essentially, it means you can dictate the speaking style using natural language instructions. What’s more, you can provide a speech reference to establish the voice timbre—the unique quality of a voice—without needing extensive training data, as mentioned in the release.

FlexiVoice is built around a Large Language Model (LLM) core. This core takes text as its primary input. It also accepts an optional natural language instruction for style control. An optional speech reference can be used to manage the timbre, the company reports. The system employs a “Progressive Post-Training (PPT) scheme” to achieve its accurate and flexible control capabilities.

Why This Matters to You

Imagine creating audio content where the AI narrator perfectly matches the mood of your story. Think of it as having a voice actor who can instantly adopt any style you describe. This system empowers content creators, podcasters, and developers to produce highly nuanced and expressive audio. It removes many previous limitations of synthetic speech.

For example, if you are producing an audiobook, you could instruct FlexiVoice to read a dramatic scene “with a suspenseful and low-pitched tone.” Then, for a lighthearted section, you could simply ask it to speak “in a cheerful and upbeat manner.” The flexibility is truly remarkable. How will you use this precise control to enhance your projects?

“FlexiVoice surpasses competing baselines and demonstrates strong capability in decoupling control factors,” the paper states. This indicates its superior performance compared to existing systems. Human evaluations further confirm its naturalness, controllability, and robustness, according to the research.

Here’s a quick look at FlexiVoice’s core capabilities:

  • Zero-Shot Voice Cloning: Replicates a voice from a minimal audio sample.
  • Natural Language Style Control: Adjusts speaking style using simple text commands.
  • Decoupled Control: Separates style, timbre, and textual content for independent adjustment.

The Surprising Finding

The most surprising aspect of FlexiVoice is its ability to disentangle control factors. The team developed a multi-objective Group Relative Policy Optimization (GRPO) to achieve this, the technical report explains. This means it can independently manage the style instruction, the reference timbre, and the textual content. Previously, these elements were often intertwined, making fine-grained control difficult. This disentanglement is crucial for truly flexible voice generation.

It challenges the common assumption that achieving highly realistic and controllable AI voices requires immense, meticulously labeled datasets for every desired style. Instead, FlexiVoice leverages an LLM core and post-training techniques. This allows for a more intuitive and adaptable approach to voice synthesis. The system’s ability to accurately follow both natural language instructions and speech references simultaneously is a significant leap, the study finds.

What Happens Next

We can expect to see early applications of FlexiVoice system emerge within the next 6 to 12 months. This could include enhanced accessibility tools, more dynamic virtual assistants, and content creation platforms. For instance, a game developer might use FlexiVoice to generate character dialogue. They could specify voices and emotional delivery without needing to record countless lines with human actors.

Actionable advice for content creators is to start experimenting with natural language prompts for voice generation. Familiarize yourself with the concept of zero-shot voice cloning. This will prepare you for the next generation of AI audio tools. The industry implications are vast, suggesting a future where synthetic speech is virtually indistinguishable from human speech. What’s more, it will offer unparalleled creative control. The team anticipates further refinements to instruction following capabilities, as detailed in the blog post.

Ready to start creating?

Create Voiceover

Transcribe Speech

Create Dialogues

Create Visuals

Clone a Voice