AutoEval Done Right: Using Synthetic Data for Model Evaluation
About
The evaluation of machine learning models using human-labeled validation data can be expensive and time-consuming. AI-labeled synthetic data can be used to decrease the number of human annotations required for this purpose in a process called autoevaluation. We suggest efficient and statistically principled algorithms for this purpose that improve sample efficiency while remaining unbiased. These algorithms increase the effective human-labeled sample size by up to 50% on experiments with GPT-4.
Pierre Boyeau, Anastasios N. Angelopoulos, Nir Yosef, Jitendra Malik, Michael I. Jordan• 2024
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Dataset-level accuracy estimation | Spider to SynSQL 2.5M | MAE9.5 | 54 | |
| Dataset-level accuracy estimation | Spider to BIRD | MAE11.6 | 54 | |
| Dataset-level accuracy estimation | WikiSQL to Spider | MAE11.1 | 54 | |
| Dataset-level accuracy estimation | SParC to CoSQL | MAE5.5 | 54 | |
| Dataset-level accuracy estimation | WikiSQL to Spider 2.0 | MAE16.3 | 54 |
Showing 5 of 5 rows