Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

TICL: Text-Embedding KNN For Speech In-Context Learning Unlocks Speech Recognition Abilities of Large Multimodal Models

About

Speech foundation models have recently demonstrated the ability to perform Speech In-Context Learning (SICL). Selecting effective in-context examples is crucial for SICL performance, yet selection methodologies remain underexplored. In this work, we propose Text-Embedding KNN for SICL (TICL), a simple pipeline that uses semantic context to enhance off-the-shelf large multimodal models' speech recognition ability without fine-tuning. Across challenging automatic speech recognition tasks, including accented English, multilingual speech, and children's speech, our method enables models to surpass zero-shot performance with up to 84.7% relative WER reduction. We conduct ablation studies to show the robustness and efficiency of our method.

Haolong Zheng, Yekaterina Yegorova, Mark Hasegawa-Johnson• 2025

Related benchmarks

TaskDatasetResultRank
Child's Automatic Speech RecognitionRSR
WER18.9
22
Speech RecognitionOGI Kids’ Speech Corpus
WER (%)8.52
9
Speech RecognitionEdmonton Narrative Norms Instrument (ENNI)
WER13.54
9
Speech RecognitionMy Science Tutor (MyST)
WER11.69
9
Showing 4 of 4 rows

Other info

Follow for update