Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Spirit LM: Interleaved Spoken and Written Language Model

About

We introduce Spirit LM, a foundation multimodal language model that freely mixes text and speech. Our model is based on a 7B pretrained text language model that we extend to the speech modality by continuously training it on text and speech units. Speech and text sequences are concatenated as a single stream of tokens, and trained with a word-level interleaving method using a small automatically-curated speech-text parallel corpus. Spirit LM comes in two versions: a Base version that uses speech phonetic units (HuBERT) and an Expressive version that models expressivity using pitch and style units in addition to the phonetic units. For both versions, the text is encoded with subword BPE tokens. The resulting model displays both the semantic abilities of text models and the expressive abilities of speech models. Additionally, we demonstrate that Spirit LM can learn new tasks in a few-shot fashion across modalities (i.e. ASR, TTS, Speech Classification). We make available model weights and inference code.

Tu Anh Nguyen, Benjamin Muller, Bokai Yu, Marta R. Costa-jussa, Maha Elbayad, Sravya Popuri, Christophe Ropers, Paul-Ambroise Duquenne, Robin Algayres, Ruslan Mavlyutov, Itai Gat, Mary Williamson, Gabriel Synnaeve, Juan Pino, Benoit Sagot, Emmanuel Dupoux• 2024

Related benchmarks

TaskDatasetResultRank
Automatic Speech RecognitionLibriSpeech (test-other)
WER11
966
Code GenerationHumanEval--
850
Multi-task Language UnderstandingMMLU
Accuracy36.9
842
Automatic Speech RecognitionLibriSpeech clean (test)
WER6
833
Language UnderstandingMMLU
Accuracy36.9
756
Mathematical ReasoningGSM8K
Accuracy (GSM8K)21.5
358
General KnowledgeMMLU
MMLU General Knowledge Accuracy36.9
170
Question AnsweringTriviaQA
Accuracy42
85
Automatic Speech RecognitionLibriSpeech Other
WER11
75
Automatic Speech RecognitionLibriSpeech Clean
WER6
57
Showing 10 of 57 rows

Other info

Code

Follow for update