Generative Spoken Dialogue Language Modeling
About
We introduce dGSLM, the first "textless" model able to generate audio samples of naturalistic spoken dialogues. It uses recent work on unsupervised spoken unit discovery coupled with a dual-tower transformer architecture with cross-attention trained on 2000 hours of two-channel raw conversational audio (Fisher dataset) without any text or labels. We show that our model is able to generate speech, laughter and other paralinguistic signals in the two channels simultaneously and reproduces more naturalistic and fluid turn-taking compared to a text-based cascaded model.
Tu Anh Nguyen, Eugene Kharitonov, Jade Copet, Yossi Adi, Wei-Ning Hsu, Ali Elkahky, Paden Tomasello, Robin Algayres, Benoit Sagot, Abdelrahman Mohamed, Emmanuel Dupoux• 2022
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Interruption Handling | Full-Duplex-Bench | GPT-4o Score0.2 | 18 | |
| Turn Taking | Full-Duplex-Bench | TOR98 | 17 | |
| Pause Handling | Full-Duplex-Bench Candor | TOR0.94 | 13 | |
| User Interruption | Bilingual Full-Duplex-Bench English | RL2.531 | 12 | |
| Backchanneling | Full-Duplex-Bench | TOR69 | 11 | |
| Pause Handling | Full-Duplex-Bench Synthetic | TOR93 | 11 | |
| Spoken Dialogue | MultiDialog 1.0 (test) | PPL1.09e+3 | 8 | |
| Overall Evaluation | Bilingual Full-Duplex-Bench English | Accuracy65.3 | 8 | |
| Duplex Dialogue Turn-Taking | Full-Duplex-Bench | Synthetic TOR for Pause Handling0.934 | 8 | |
| Pause Handling | Bilingual Full-Duplex-Bench English | TOR93.5 | 6 |
Showing 10 of 16 rows