OmniVoice: Towards Omnilingual Zero-Shot Text-to-Speech with Diffusion Language Models
About
We present OmniVoice, a massive multilingual zero-shot text-to-speech (TTS) model that scales to over 600 languages. At its core is a novel diffusion language model-style discrete non-autoregressive (NAR) architecture. Unlike conventional discrete NAR models that suffer from performance bottlenecks in complex two-stage (text-to-semantic-to-acoustic) pipelines, OmniVoice directly maps text to multi-codebook acoustic tokens. This simplified approach is facilitated by two key technical innovations: (1) a full-codebook random masking strategy for efficient training, and (2) initialization from a pre-trained LLM to ensure superior intelligibility. By leveraging a 581k-hour multilingual dataset curated entirely from open-source data, OmniVoice achieves the broadest language coverage to date and delivers state-of-the-art performance across Chinese, English, and diverse multilingual benchmarks. Our code and pre-trained models are publicly available at https://github.com/k2-fsa/OmniVoice.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Text-to-Speech | Seed-TTS en (test) | WER1.6 | 90 | |
| Text-to-Speech | MiniMax Multilingual 24 (test) | WER0.874 | 75 | |
| Text-to-Speech | Seed-TTS zh (test) | WER0.0084 | 65 | |
| Text-to-Speech | LibriSpeech PC clean (test) | WER1.3 | 31 | |
| Text-to-Speech | Chinese and English (test) | CMOS0.44 | 6 | |
| Multilingual Text-to-Speech | FLEURS-Multilingual-102 (test) | Average Character Error Rate (CER)4 | 2 |