Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

DnA-Eval: Enhancing Large Language Model Evaluation through Decomposition and Aggregation

About

The acceleration of Large Language Models (LLMs) research has opened up new possibilities for evaluating generated texts. They serve as scalable and economical evaluators, but the question of how reliable these evaluators are has emerged as a crucial research question. Prior research efforts in the meta-evaluation of LLMs as judges limit the prompting of an LLM to a single use to obtain a final evaluation decision. They then compute the agreement between LLMs' outputs and human labels. This lacks interpretability in understanding the evaluation capability of LLMs. In light of this challenge, we propose Decompose and Aggregate, which breaks down the evaluation process into different stages based on pedagogical practices. Our experiments illustrate that it not only provides a more interpretable window for how well LLMs evaluate, but also leads to improvements up to 39.6% for different LLMs on a variety of meta-evaluation benchmarks.

Minzhi Li, Zhengyuan Liu, Shumin Deng, Shafiq Joty, Nancy F. Chen, Min-Yen Kan• 2024

Related benchmarks

TaskDatasetResultRank
Human-Metric CorrelationRealHumanEval (Out-of-Distribution)
Kendall's Tau0.152
9
Human-Metric CorrelationCoGym (Out-of-Distribution)
Kendall's Tau0.353
9
Human-Metric CorrelationEvalGen Out-of-Distribution
Kendall's Tau0.232
9
Human-Metric CorrelationHelpSteer2 (In-Distribution)
Kendall's Tau0.26
9
Human-Metric CorrelationSimpEval In-Distribution
Kendall's Tau0.234
9
Showing 5 of 5 rows

Other info

Follow for update