Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SyncVSR: Data-Efficient Visual Speech Recognition with End-to-End Crossmodal Audio Token Synchronization

About

Visual Speech Recognition (VSR) stands at the intersection of computer vision and speech recognition, aiming to interpret spoken content from visual cues. A prominent challenge in VSR is the presence of homophenes-visually similar lip gestures that represent different phonemes. Prior approaches have sought to distinguish fine-grained visemes by aligning visual and auditory semantics, but often fell short of full synchronization. To address this, we present SyncVSR, an end-to-end learning framework that leverages quantized audio for frame-level crossmodal supervision. By integrating a projection layer that synchronizes visual representation with acoustic data, our encoder learns to generate discrete audio tokens from a video sequence in a non-autoregressive manner. SyncVSR shows versatility across tasks, languages, and modalities at the cost of a forward pass. Our empirical evaluations show that it not only achieves state-of-the-art results but also reduces data usage by up to ninefold.

Young Jin Ahn, Jungwoo Park, Sangha Park, Jonghyun Choi, Kee-Eung Kim• 2024

Related benchmarks

TaskDatasetResultRank
Visual Speech RecognitionLRS3
WER0.215
59
Visual Speech RecognitionLRS2
Mean WER16.5
45
Lip-readingLRW 1.0 (test)--
37
Visual Speech RecognitionLRW
Top-1 Accuracy80.3
5
Word-level Visual Speech RecognitionCAS-VSR-W1K Chinese (test)--
5
Showing 5 of 5 rows

Other info

Follow for update