Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Latent Flow Matching for Expressive Singing Voice Synthesis

About

Conditional variational autoencoder (cVAE)-based singing voice synthesis provides efficient inference and strong audio quality by learning a score-conditioned prior and a recording-conditioned posterior latent space. However, because synthesis relies on prior samples while training uses posterior latents inferred from real recordings, imperfect distribution matching can cause a prior-posterior mismatch that degrades fine-grained expressiveness such as vibrato and micro-prosody. We propose FM-Singer, which introduces conditional flow matching (CFM) in latent space to learn a continuous vector field transporting prior latents toward posterior latents along an optimal-transport-inspired path. At inference time, the learned latent flow refines a prior sample by solving an ordinary differential equation (ODE) before waveform generation, improving expressiveness while preserving the efficiency of parallel decoding. Experiments on Korean and Chinese singing datasets demonstrate consistent improvements over strong baselines, including lower mel-cepstral distortion and fundamental-frequency error and higher perceptual scores on the Korean dataset. Code, pretrained checkpoints, and audio demos are available at https://github.com/alsgur9368/FM-Singer

Minhyeok Yun, Yong-Hoon Choi• 2026

Related benchmarks

TaskDatasetResultRank
Singing Voice SynthesisOpencpop
F0 RMSE25.2
4
Singing Voice SynthesisKorean singing voice dataset (test)
MOS4.039
4
Showing 2 of 2 rows

Other info

Follow for update