Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

BERTScore: Evaluating Text Generation with BERT

About

We propose BERTScore, an automatic evaluation metric for text generation. Analogously to common metrics, BERTScore computes a similarity score for each token in the candidate sentence with each token in the reference sentence. However, instead of exact matches, we compute token similarity using contextual embeddings. We evaluate using the outputs of 363 machine translation and image captioning systems. BERTScore correlates better with human judgments and provides stronger model selection performance than existing metrics. Finally, we use an adversarial paraphrase detection task to show that BERTScore is more robust to challenging examples when compared to existing metrics.

Tianyi Zhang, Varsha Kishore, Felix Wu, Kilian Q. Weinberger, Yoav Artzi• 2019

Related benchmarks

TaskDatasetResultRank
Image Captioning EvaluationComposite
Kendall-c Tau_c30.1
131
Image Captioning EvaluationFlickr8K-CF
Kendall-b Correlation (tau_b)38.5
99
Image Captioning EvaluationFlickr8k Expert
Kendall Tau-c (tau_c)46.7
82
Image Captioning EvaluationFlickr8K Expert (test)
Kendall tau_c39.2
76
Image Captioning EvaluationPascal-50S (test)
HC65.4
66
Image Captioning EvaluationFlickr8K-CF (test)
Kendall tau_b22.8
65
Factual Consistency EvaluationSummaC
CGS63.1
52
General Instruction FollowingArena Hard
Score0.645
46
Metrics correlation with human judgmentWebNLG challenge 2017
Spearman Correlation (rho)0.81
45
Summarization EvaluationSummEval
Avg Spearman Rho0.225
45
Showing 10 of 197 rows
...

Other info

Follow for update