Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

PLR: Plackett-Luce for Reordering In-Context Learning Examples

About

In-context learning (ICL) adapts large language models by conditioning on a small set of ICL examples, avoiding costly parameter updates. Among other factors, performance is often highly sensitive to the ordering of the examples. However, exhaustive search over the $n!$ possible orderings is infeasible. Therefore more efficient ordering methods use model confidence measures (e.g., label-probability entropy) over label sets or take a direct approach to finding the best ordering. We propose PLR, a probabilistic approach to in-context example ordering that replaces discrete ordering search with learning a probability distribution over orderings with the Plackett-Luce model. PLR models orderings using a Plackett-Luce distribution and iteratively updates its parameters to concentrate probability mass on high-performing orderings under a task-level metric. Candidate orderings are sampled efficiently via a Gumbel perturb-and-sort procedure. Experiments on multiple classification benchmarks show that PLR consistently improves few-shot accuracy for $k \in \{4, 8, 16, 32\}$ examples, and we further demonstrate gains on mathematical reasoning tasks where label-based ordering methods are not applicable. Our code is available at https://github.com/Batorskq/PLR.

Pawel Batorski, Paul Swoboda• 2026

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K (test)
Accuracy42.85
900
Text ClassificationMR (test)
Accuracy93.13
148
Subjectivity ClassificationSubj (test)
Accuracy93.59
127
Text ClassificationTREC (test)
Accuracy70.63
115
Text ClassificationSST-5 (test)
Accuracy55.96
60
Mathematical ReasoningDeepMath
Accuracy46.36
30
Text ClassificationNews (test)
Accuracy86.31
2
Showing 7 of 7 rows

Other info

Follow for update