Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

D-GEN: Automatic Distractor Generation and Evaluation for Reliable Assessment of Generative Model

About

Evaluating generative models with open-ended generation is challenging due to inconsistencies in response formats. Multiple-choice (MC) evaluation mitigates this issue, but generating high-quality distractors is time-consuming and labor-intensive. We introduce D-GEN, the first open-source distractor generator model that transforms open-ended data into an MC format. To evaluate distractor quality, we propose two novel methods: (1) ranking alignment, ensuring generated distractors retain the discriminatory power of ground-truth distractors, and (2) entropy analysis, comparing model confidence distributions. Our results show that D-GEN preserves ranking consistency (Spearman's rho 0.99, Kendall's tau 0.94) and closely matches the entropy distribution of ground-truth distractors. Human evaluation further confirms the fluency, coherence, distractiveness, and incorrectness. Our work advances robust and efficient distractor generation with automated evaluation, setting a new standard for MC evaluation.

Grace Byun, Jinho D. Choi• 2025

Related benchmarks

TaskDatasetResultRank
Distractor GenerationD-GEN
Fluency4.98
1
Distractor GenerationD-GEN Commonsense Reasoning
Fluency4.97
1
Distractor GenerationD-GEN RC + CS
Fluency4.99
1
Distractor GenerationD-GEN Translation
Fluency4.91
1
Distractor GenerationD-GEN Summarization
Fluency4.96
1
Distractor GenerationD-GEN Struct-to-Text
Fluency4.88
1
Distractor GenerationD-GEN Mathematics
Fluency5
1
Showing 7 of 7 rows

Other info

Code

Follow for update