Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

GPTScore: Evaluate as You Desire

About

Generative Artificial Intelligence (AI) has enabled the development of sophisticated models that are capable of producing high-caliber text, images, and other outputs through the utilization of large pre-trained models. Nevertheless, assessing the quality of the generation is an even more arduous task than the generation itself, and this issue has not been given adequate consideration recently. This paper proposes a novel evaluation framework, GPTScore, which utilizes the emergent abilities (e.g., zero-shot instruction) of generative pre-trained models to score generated texts. There are 19 pre-trained models explored in this paper, ranging in size from 80M (e.g., FLAN-T5-small) to 175B (e.g., GPT3). Experimental results on four text generation tasks, 22 evaluation aspects, and corresponding 37 datasets demonstrate that this approach can effectively allow us to achieve what one desires to evaluate for texts simply by natural language instructions. This nature helps us overcome several long-standing challenges in text evaluation--how to achieve customized, multi-faceted evaluation without the need for annotated samples. We make our code publicly available at https://github.com/jinlanfu/GPTScore.

Jinlan Fu, See-Kiong Ng, Zhengbao Jiang, Pengfei Liu• 2023

Related benchmarks

TaskDatasetResultRank
Summarization EvaluationSummEval
Avg Spearman Rho0.394
40
Factual Consistency EvaluationQAGS XSUM
Spearman Correlation22
39
Factual Consistency EvaluationSummEval
Spearman Correlation0.475
36
Quantitative evaluation of LLM feedback against human gold standards50 SOC analysis reports (test)
Spearman Correlation (ρ)0.65
30
Dialogue Evaluation Human CorrelationTopical-Chat
Naturalness Pearson (r)0.353
26
Meta-evaluationSummEval--
10
Patent Quality EvaluationPap2Pat EvalGold N=146 (test)
TCF38
8
Showing 7 of 7 rows

Other info

Follow for update