Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Towards General Text Embeddings with Multi-stage Contrastive Learning

About

We present GTE, a general-purpose text embedding model trained with multi-stage contrastive learning. In line with recent advancements in unifying various NLP tasks into a single format, we train a unified text embedding model by employing contrastive learning over a diverse mixture of datasets from multiple sources. By significantly increasing the number of training data during both unsupervised pre-training and supervised fine-tuning stages, we achieve substantial performance gains over existing embedding models. Notably, even with a relatively modest parameter count of 110M, GTE$_\text{base}$ outperforms the black-box embedding API provided by OpenAI and even surpasses 10x larger text embedding models on the massive text embedding benchmark. Furthermore, without additional fine-tuning on each programming language individually, our model outperforms previous best code retrievers of similar size by treating code as text. In summary, our model achieves impressive results by effectively harnessing multi-stage contrastive learning, offering a powerful and efficient text embedding model with broad applicability across various NLP and code-related tasks.

Zehan Li, Xin Zhang, Yanzhao Zhang, Dingkun Long, Pengjun Xie, Meishan Zhang• 2023

Related benchmarks

TaskDatasetResultRank
Multi-hop QAHotpotQA
Exact Match58.6
76
Text EmbeddingMTEB English v2
Mean Score67.2
68
Multi-hop QAMuSiQue
EM30.6
65
RetrievalNatural Questions (test)
Top-5 Recall74.3
62
Information RetrievalBEIR--
62
Natural Language UnderstandingGLUE (test val)
MRPC Accuracy92.1
59
Sentence Embedding EvaluationMTEB (test)
Classification Score86.58
55
Tool CallingAPI-Bank L-1--
46
Information RetrievalBRIGHT 1.0 (test)
nDCG@10 (Avg)22.8
35
Multi-hop QA Retrieval2WikiMultiHopQA (test)
R@574.8
33
Showing 10 of 104 rows
...

Other info

Follow for update