MelNet: A Generative Model for Audio in the Frequency Domain
About
Capturing high-level structure in audio waveforms is challenging because a single second of audio spans tens of thousands of timesteps. While long-range dependencies are difficult to model directly in the time domain, we show that they can be more tractably modelled in two-dimensional time-frequency representations such as spectrograms. By leveraging this representational advantage, in conjunction with a highly expressive probabilistic model and a multiscale generation procedure, we design a model capable of generating high-fidelity audio samples which capture structure at timescales that time-domain models have yet to achieve. We apply our model to a variety of audio generation tasks, including unconditional speech generation, music generation, and text-to-speech synthesis---showing improvements over previous approaches in both density estimates and human judgments.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Unconditional Music Generation | MAESTRO | Selection Rate95.8 | 4 | |
| Unconditional Audio Generation | Blizzard 2013 | Selection Rate100 | 2 | |
| Unconditional Audio Generation | VoxCeleb2 | Selection Rate1.00e+4 | 2 |