Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning to Retrieve In-Context Examples for Large Language Models

About

Large language models (LLMs) have demonstrated their ability to learn in-context, allowing them to perform various tasks based on a few input-output examples. However, the effectiveness of in-context learning is heavily reliant on the quality of the selected examples. In this paper, we propose a novel framework to iteratively train dense retrievers that can identify high-quality in-context examples for LLMs. Our framework initially trains a reward model based on LLM feedback to evaluate the quality of candidate examples, followed by knowledge distillation to train a bi-encoder based dense retriever. Our experiments on a suite of $30$ tasks demonstrate that our framework significantly enhances in-context learning performance. Furthermore, we show the generalization ability of our framework to unseen tasks during training. An in-depth analysis reveals that our model improves performance by retrieving examples with similar patterns, and the gains are consistent across LLMs of varying sizes. The code and data are available at https://github.com/microsoft/LMOps/tree/main/llm_retriever .

Liang Wang, Nan Yang, Furu Wei• 2023

Related benchmarks

TaskDatasetResultRank
Commonsense ReasoningHellaSwag
Accuracy74.6
1460
Natural Language InferenceRTE
Accuracy61.7
367
Reading ComprehensionBoolQ
Accuracy74.9
219
Natural Language InferenceSNLI
Accuracy80
174
Topic ClassificationAG-News
Accuracy92.4
173
Sentiment AnalysisSST-2
Accuracy93.4
156
Common Sense ReasoningCOPA
Accuracy85
138
Sentiment AnalysisSST-2 (test)
Accuracy94.3
136
Natural language generationE2E (test)
ROUGE-L56.4
79
Paraphrase IdentificationQQP
Accuracy80.9
78
Showing 10 of 25 rows

Other info

Follow for update