Meta-Task Prompting Elicits Embeddings from Large Language Models
About
We introduce a new unsupervised text embedding method, Meta-Task Prompting with Explicit One-Word Limitation (MetaEOL), for generating high-quality sentence embeddings from Large Language Models (LLMs) without the need for model fine-tuning. Leveraging meta-task prompting, MetaEOL guides LLMs to produce embeddings through a series of carefully designed prompts that address multiple representational aspects. Our comprehensive experiments demonstrate that embeddings averaged from various meta-tasks are versatile embeddings that yield competitive performance on Semantic Textual Similarity (STS) benchmarks and excel in downstream tasks, surpassing contrastive-trained models. Our findings suggest a new scaling law, offering a versatile and resource-efficient approach for embedding generation across diverse scenarios.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Semantic Textual Similarity | STS tasks (STS12, STS13, STS14, STS15, STS16, STS-B, SICK-R) various (test) | STS12 Score65.1 | 393 | |
| Sentence Embedding Evaluation | MTEB (test) | -- | 48 | |
| Clustering | MTEB Clustering | Bior Score30.95 | 23 | |
| Transfer Learning | SentEval Transfer tasks (test) | MR90.93 | 23 | |
| Information Retrieval | MTEB Retrieval | SciF Score40.59 | 5 | |
| Semantic Textual Similarity | MTEB (test) | -- | 4 |