WhisperX: Time-Accurate Speech Transcription of Long-Form Audio
About
Large-scale, weakly-supervised speech recognition models, such as Whisper, have demonstrated impressive results on speech recognition across domains and languages. However, their application to long audio transcription via buffered or sliding window approaches is prone to drifting, hallucination & repetition; and prohibits batched transcription due to their sequential nature. Further, timestamps corresponding each utterance are prone to inaccuracies and word-level timestamps are not available out-of-the-box. To overcome these challenges, we present WhisperX, a time-accurate speech recognition system with word-level timestamps utilising voice activity detection and forced phoneme alignment. In doing so, we demonstrate state-of-the-art performance on long-form transcription and word segmentation benchmarks. Additionally, we show that pre-segmenting audio with our proposed VAD Cut & Merge strategy improves transcription quality and enables a twelve-fold transcription speedup via batched inference.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Forced Alignment | MFA-Labeled Raw (test) | AAS Latency (Avg)133.2 | 8 | |
| Forced Alignment | Human-Labeled (test) | Avg. RTF0.0113 | 4 | |
| Forced Alignment | MFA-labeled Long-form (test) | Average Alignment Value2.71e+3 | 4 | |
| Speaker Diarization | StoryGen Eval | tcpWER55.9 | 3 | |
| Automatic Speech Recognition | Russian long audio lectures (test) | WER0.1683 | 2 |