Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SUPERT: Towards New Frontiers in Unsupervised Evaluation Metrics for Multi-Document Summarization

About

We study unsupervised multi-document summarization evaluation metrics, which require neither human-written reference summaries nor human annotations (e.g. preferences, ratings, etc.). We propose SUPERT, which rates the quality of a summary by measuring its semantic similarity with a pseudo reference summary, i.e. selected salient sentences from the source documents, using contextualized embeddings and soft token alignment techniques. Compared to the state-of-the-art unsupervised evaluation metrics, SUPERT correlates better with human ratings by 18-39%. Furthermore, we use SUPERT as rewards to guide a neural-based reinforcement learning summarizer, yielding favorable performance compared to the state-of-the-art unsupervised summarizers. All source code is available at https://github.com/yg211/acl20-ref-free-eval.

Yang Gao, Wei Zhao, Steffen Eger• 2020

Related benchmarks

TaskDatasetResultRank
Summarization EvaluationSummEval 1.0 (test)
Coherence (Spearman rho)0.2165
21
Summarization EvaluationSummEval Relevance Domain
Corr.0.91
8
Showing 2 of 2 rows

Other info

Follow for update