Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Active Example Selection for In-Context Learning

About

With a handful of demonstration examples, large-scale language models show strong capability to perform various tasks by in-context learning from these examples, without any fine-tuning. We demonstrate that in-context learning performance can be highly unstable across samples of examples, indicating the idiosyncrasies of how language models acquire information. We formulate example selection for in-context learning as a sequential decision problem, and propose a reinforcement learning algorithm for identifying generalizable policies to select demonstration examples. For GPT-2, our learned policies demonstrate strong abilities of generalizing to unseen tasks in training, with a $5.8\%$ improvement on average. Examples selected from our learned policies can even achieve a small improvement on GPT-3 Ada. However, the improvement diminishes on larger GPT-3 models, suggesting emerging capabilities of large language models.

Yiming Zhang, Shi Feng, Chenhao Tan• 2022

Related benchmarks

TaskDatasetResultRank
Commonsense ReasoningHellaSwag
Accuracy86.83
1460
Natural Language InferenceRTE
Accuracy47.5
367
Natural Language InferenceSNLI
Accuracy35
174
Intent ClassificationBanking77 (test)
Accuracy84.2
151
Commonsense Question AnsweringCommonsenseQA
Accuracy87.55
81
Sentiment AnalysisSST-5
Accuracy43.69
47
Natural Language InferenceQNLI
Accuracy61.5
42
Natural Language InferenceMNLI
Accuracy70.92
36
Natural Language InferenceMNLI mm
Accuracy29.5
30
Paraphrase DetectionPAWS
Accuracy51.7
24
Showing 10 of 21 rows

Other info

Follow for update