Freeze-Omni: A Smart and Low Latency Speech-to-speech Dialogue Model with Frozen LLM
About
Rapidly developing large language models (LLMs) have brought tremendous intelligent applications. Especially, the GPT-4o's excellent duplex speech interaction ability has brought impressive experience to users. Researchers have recently proposed several multi-modal LLMs in this direction that can achieve user-agent speech-to-speech conversations. This paper proposes a novel speech-text multimodal LLM architecture called Freeze-Omni. Our main contribution is that the speech input and output modalities can be easily connected to a textual LLM while keeping the LLM's parameters frozen throughout the training process. We design a three-stage training strategy for modeling both the speech input and output, enabling Freeze-Omni to obtain speech-to-speech conversation ability using text-speech paired data (such as ASR and TTS data) and only 60,000 multi-round text Q&A data on 8 GPUs. Moreover, we can effectively ensure that the intelligence of the Freeze-Omni in the speech modality is at the same level compared with that in the text modality of its backbone LLM, while achieving low latency end-to-end spoken response. In addition, we also designed a method to achieve duplex dialogue ability through multi-task training, giving Freeze-Omni a more natural style of dialogue ability between users and agents. In summary, Freeze-Omni holds great potential to conduct speech-to-speech dialogue based on a multimodal LLM under the condition of a frozen LLM, avoiding the catastrophic forgetting problem caused by limited data and training resources.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Speech-to-Text Question-Answering | WebQ | Accuracy54.1 | 23 | |
| Speech-to-Text Question-Answering | LlamaQ | Accuracy78.67 | 23 | |
| Speech-to-Text Question-Answering | TriviaQA | Accuracy51.15 | 23 | |
| General Audio Understanding | VoiceBench | AlpacaEval Score4.03 | 19 | |
| Interruption Handling | Full-Duplex-Bench | GPT-4o Score3.62 | 18 | |
| Factuality Evaluation | TriviaQA | Response Accuracy53.9 | 18 | |
| Factuality Evaluation | WebQ | Accuracy (Response)44.7 | 18 | |
| Factuality Evaluation | LlamaQ | Response Accuracy72 | 18 | |
| Turn Taking | Full-Duplex-Bench | TOR34 | 17 | |
| Speech-to-Speech Question-Answering | Llama Questions | Accuracy58.67 | 15 |