Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

WaveFM: A High-Fidelity and Efficient Vocoder Based on Flow Matching

About

Flow matching offers a robust and stable approach to training diffusion models. However, directly applying flow matching to neural vocoders can result in subpar audio quality. In this work, we present WaveFM, a reparameterized flow matching model for mel-spectrogram conditioned speech synthesis, designed to enhance both sample quality and generation speed for diffusion vocoders. Since mel-spectrograms represent the energy distribution of waveforms, WaveFM adopts a mel-conditioned prior distribution instead of a standard Gaussian prior to minimize unnecessary transportation costs during synthesis. Moreover, while most diffusion vocoders rely on a single loss function, we argue that incorporating auxiliary losses, including a refined multi-resolution STFT loss, can further improve audio quality. To speed up inference without degrading sample quality significantly, we introduce a tailored consistency distillation method for WaveFM. Experiment results demonstrate that our model achieves superior performance in both quality and efficiency compared to previous diffusion vocoders, while enabling waveform generation in a single inference step.

Tianze Luo, Xingchen Miao, Wenbo Duan• 2025

Related benchmarks

TaskDatasetResultRank
Speech SynthesisLibriTTS (test)--
17
Text-to-SpeechLibriSpeech clean PC (test)--
17
Zero-shot Text-to-SpeechLibriSpeech PC clean (test)
WER2.01
12
Showing 3 of 3 rows

Other info

Follow for update