Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Smoothie: Label Free Language Model Routing

About

Large language models (LLMs) are increasingly used in applications where LLM inputs may span many different tasks. Recent work has found that the choice of LLM is consequential, and different LLMs may be good for different input samples. Prior approaches have thus explored how engineers might select an LLM to use for each sample (i.e. routing). While existing routing methods mostly require training auxiliary models on human-annotated data, our work explores whether it is possible to perform unsupervised routing. We propose Smoothie, a weak supervision-inspired routing approach that requires no labeled data. Given a set of outputs from different LLMs, Smoothie constructs a latent variable graphical model over embedding representations of observable LLM outputs and unknown "true" outputs. Using this graphical model, we estimate sample-dependent quality scores for each LLM, and route each sample to the LLM with the highest corresponding score. We find that Smoothie's LLM quality-scores correlate with ground-truth model quality (correctly identifying the optimal model on 9/14 tasks), and that Smoothie outperforms baselines for routing by up to 10 points accuracy.

Neel Guha, Mayee F. Chen, Trevor Chow, Ishan S. Khare, Christopher R\'e• 2024

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K
Accuracy (GSM8K)37.5
358
SummarizationXSum (test)
ROUGE-28.4
231
Question AnsweringTriviaQA
Accuracy68.7
210
Arithmetic ReasoningGSM8K
Accuracy91.5
155
Instruction FollowingAlpacaEval
Win Rate34.5
125
Question AnsweringTriviaQA (test)
Accuracy68.3
121
Question AnsweringSQuAD (test)--
111
SummarizationXsum
ROUGE-212.8
108
Question AnsweringSQuAD
Exact Match63.1
50
Data-to-text generationWebNLG (test)--
39
Showing 10 of 33 rows

Other info

Code

Follow for update