Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

VATLM: Visual-Audio-Text Pre-Training with Unified Masked Prediction for Speech Representation Learning

About

Although speech is a simple and effective way for humans to communicate with the outside world, a more realistic speech interaction contains multimodal information, e.g., vision, text. How to design a unified framework to integrate different modal information and leverage different resources (e.g., visual-audio pairs, audio-text pairs, unlabeled speech, and unlabeled text) to facilitate speech representation learning was not well explored. In this paper, we propose a unified cross-modal representation learning framework VATLM (Visual-Audio-Text Language Model). The proposed VATLM employs a unified backbone network to model the modality-independent information and utilizes three simple modality-dependent modules to preprocess visual, speech, and text inputs. In order to integrate these three modalities into one shared semantic space, VATLM is optimized with a masked prediction task of unified tokens, given by our proposed unified tokenizer. We evaluate the pre-trained VATLM on audio-visual related downstream tasks, including audio-visual speech recognition (AVSR), visual speech recognition (VSR) tasks. Results show that the proposed VATLM outperforms previous the state-of-the-art models, such as audio-visual pre-trained AV-HuBERT model, and analysis also demonstrates that VATLM is capable of aligning different modalities into the same space. To facilitate future research, we release the code and pre-trained models at https://aka.ms/vatlm.

Qiushi Zhu, Long Zhou, Ziqiang Zhang, Shujie Liu, Binxing Jiao, Jie Zhang, Lirong Dai, Daxin Jiang, Jinyu Li, Furu Wei• 2022

Related benchmarks

TaskDatasetResultRank
Visual Speech RecognitionLRS3 (test)
WER2.7
159
Visual Speech RecognitionLRS3 High-Resource, 433h labelled v1 (test)
WER0.012
80
Visual Speech RecognitionLRS3
WER0.262
59
Visual Speech RecognitionLRS2
Mean WER24.3
45
Visual Speech RecognitionLRS3 Low-Resource 30h labelled v1 (test)
WER0.027
34
Speech RecognitionLRS3 high-resource
WER (V)28.4
18
Showing 6 of 6 rows

Other info

Follow for update