Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

SONAR: Sentence-Level Multimodal and Language-Agnostic Representations

About

We introduce SONAR, a new multilingual and multimodal fixed-size sentence embedding space. Our single text encoder, covering 200 languages, substantially outperforms existing sentence embeddings such as LASER3 and LabSE on the xsim and xsim++ multilingual similarity search tasks. Speech segments can be embedded in the same SONAR embedding space using language-specific speech encoders trained in a teacher-student setting on speech transcription data. Our encoders outperform existing speech encoders on similarity search tasks. We also provide a text decoder for 200 languages, which allows us to perform text-to-text and speech-to-text machine translation, including for zero-shot language and modality combinations. Our text-to-text results are competitive compared to the state-of-the-art NLLB~1B model, despite the fixed-size bottleneck representation. Our zero-shot speech-to-text translation results compare favorably with strong supervised baselines such as Whisper.

Paul-Ambroise Duquenne, Holger Schwenk, Beno\^it Sagot• 2023

Related benchmarks

TaskDatasetResultRank
Multimodal Sentiment AnalysisCMU-MOSI--
144
Speech Emotion RecognitionRAVDESS
Unweighted Accuracy10.8
43
Speech Emotion RecognitionMELD--
19
Depression DetectionDAIC-WOZ
Weighted F1-score64.3
8
Speech Emotion RecognitionIEMOCAP 4
Weighted F1-score59.4
8
Speech Emotion RecognitionIEMOCAP-6
Weighted F143.5
8
Cross-modal and cross-lingual retrievalFLEURS (test)
Avg xSIM++ (26)14.3
7
Machine TranslationFLORES X→Eng (devtest)
COMET Score (Low)0.851
6
Cross-lingual similarity searchFLORES X→Eng (devtest)
xSIM++ Low13.1
5
Multilingual ClassificationMTEB
Average Accuracy63.02
4
Showing 10 of 20 rows

Other info

Follow for update