StarGANv2-VC: A Diverse, Unsupervised, Non-parallel Framework for Natural-Sounding Voice Conversion
About
We present an unsupervised non-parallel many-to-many voice conversion (VC) method using a generative adversarial network (GAN) called StarGAN v2. Using a combination of adversarial source classifier loss and perceptual loss, our model significantly outperforms previous VC models. Although our model is trained only with 20 English speakers, it generalizes to a variety of voice conversion tasks, such as any-to-many, cross-lingual, and singing conversion. Using a style encoder, our framework can also convert plain reading speech into stylistic speech, such as emotional and falsetto speech. Subjective and objective evaluation experiments on a non-parallel many-to-many voice conversion task revealed that our model produces natural sounding voices, close to the sound quality of state-of-the-art text-to-speech (TTS) based voice conversion methods without the need for text labels. Moreover, our model is completely convolutional and with a faster-than-real-time vocoder such as Parallel WaveGAN can perform real-time voice conversion.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Emotion Transfer | ESD, TIMIT, and CREMA-D Evaluation Suite (test) | SSST5.72 | 20 | |
| Rhythm Transfer | ESD, TIMIT, and CREMA-D Evaluation Suite (test) | SSST54 | 10 | |
| Emotion Voice Conversion | Emotion Speech Dataset (ESD) (test) | Speaker Centered Similarity (Spk-Cent SIM)0.565 | 6 | |
| Speaker Preservation | Speech Style Transfer (test) | Speaker Similarity (Source->Target)0.76 | 5 | |
| Content Preservation | Speech Style Transfer (test) | SSST5.72 | 5 |