Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

WavSLM: Single-Stream Speech Language Modeling via WavLM Distillation

About

Large language models show that simple autoregressive training can yield scalable and coherent generation, but extending this paradigm to speech remains challenging due to the entanglement of semantic and acoustic information. Most existing speech language models rely on text supervision, hierarchical token streams, or complex hybrid architectures, departing from the single-stream generative pretraining paradigm that has proven effective in text. In this work, we introduce WavSLM, a speech language model trained by quantizing and distilling self-supervised WavLM representations into a single codebook and optimizing an autoregressive next-chunk prediction objective. WavSLM jointly models semantic and acoustic information within a single token stream without text supervision or text pretraining. Despite its simplicity, it achieves competitive performance on consistency benchmarks and speech generation while using fewer parameters, less training data, and supporting streaming inference. Demo samples are available at https://lucadellalib.github.io/wavslm-web/.

Luca Della Libera, Cem Subakan, Mirco Ravanelli• 2026

Related benchmarks

TaskDatasetResultRank
Semantic and linguistic knowledge evaluationZeroSpeech
sBLiMP Score54.6
20
Acoustic and paralinguistic modelingSALMon
Acoustic Consistency (Sentence)75.5
19
Discourse-level coherence evaluationTopic Story-Cloze (tSC)
tSC Score63.3
19
Speech GenerationLibrispeech (test-clean)
Speaker Similarity0.918
11
Showing 4 of 4 rows

Other info

Follow for update