Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

MEXMA: Token-level objectives improve sentence representations

About

Current pre-trained cross-lingual sentence encoders approaches use sentence-level objectives only. This can lead to loss of information, especially for tokens, which then degrades the sentence representation. We propose MEXMA, a novel approach that integrates both sentence-level and token-level objectives. The sentence representation in one language is used to predict masked tokens in another language, with both the sentence representation and all tokens directly updating the encoder. We show that adding token-level objectives greatly improves the sentence representation quality across several tasks. Our approach outperforms current pre-trained cross-lingual sentence encoders on bi-text mining as well as several downstream tasks. We also analyse the information encoded in our tokens, and how the sentence representation is built from them.

Jo\~ao Maria Janeiro, Benjamin Piwowarski, Patrick Gallinari, Lo\"ic Barrault• 2024

Related benchmarks

TaskDatasetResultRank
Word Sense DisambiguationWiC--
87
Sentence Embedding EvaluationMTEB (test)
Classification Score65.197
55
Named Entity RecognitionPAN-X--
16
Slot FillingMASSIVE Slotfill
F139.6
14
Code-to-Code RetrievalXLCoST
C Retrieval Score18.9
9
X-Eng cross-lingual similarity searchFLORES200 Full 201
xSim Score15.9
8
X-Eng cross-lingual similarity searchBOUQuET 177
xSim Score39.3
8
X-Eng cross-lingual similarity searchAfroMT 38
xsim58.7
8
X-Eng cross-lingual similarity searchFLORES+ 212
xsim17.5
8
X-Eng cross-lingual similarity searchBIBLE 1,560
xsim70.2
8
Showing 10 of 28 rows

Other info

Follow for update