Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

vq-wav2vec: Self-Supervised Learning of Discrete Speech Representations

About

We propose vq-wav2vec to learn discrete representations of audio segments through a wav2vec-style self-supervised context prediction task. The algorithm uses either a gumbel softmax or online k-means clustering to quantize the dense representations. Discretization enables the direct application of algorithms from the NLP community which require discrete inputs. Experiments show that BERT pre-training achieves a new state of the art on TIMIT phoneme classification and WSJ speech recognition.

Alexei Baevski, Steffen Schneider, Michael Auli• 2019

Related benchmarks

TaskDatasetResultRank
Automatic Speech RecognitionLibriSpeech clean (test)
WER6.2
1156
Automatic Speech RecognitionLibriSpeech (test-other)
WER18.2
1151
Automatic Speech RecognitionLibriSpeech (dev-other)
WER15.5
462
Automatic Speech RecognitionLibriSpeech (dev-clean)
WER (%)5.6
340
Speech RecognitionWSJ (92-eval)
WER8.57
131
Speech RecognitionWSJ nov93 (dev)
WER4.46
52
Image ReconstructionCelebA-HQ (test)
FID (Reconstruction)12.03
50
Semantic Image SynthesisADE20K (val)
FID37.51
47
Speech RecognitionWSJ nov92 (test)
WER2.34
34
Phoneme RecognitionTIMIT (test)
PER11.4
31
Showing 10 of 21 rows

Other info

Code

Follow for update