Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

A Hierarchical Network for Abstractive Meeting Summarization with Cross-Domain Pretraining

About

With the abundance of automatic meeting transcripts, meeting summarization is of great interest to both participants and other parties. Traditional methods of summarizing meetings depend on complex multi-step pipelines that make joint optimization intractable. Meanwhile, there are a handful of deep neural models for text summarization and dialogue systems. However, the semantic structure and styles of meeting transcripts are quite different from articles and conversations. In this paper, we propose a novel abstractive summary network that adapts to the meeting scenario. We design a hierarchical structure to accommodate long meeting transcripts and a role vector to depict the difference among speakers. Furthermore, due to the inadequacy of meeting summary data, we pretrain the model on large-scale news summary data. Empirical results show that our model outperforms previous approaches in both automatic metrics and human evaluation. For example, on ICSI dataset, the ROUGE-1 score increases from 34.66% to 46.28%.

Chenguang Zhu, Ruochen Xu, Michael Zeng, Xuedong Huang• 2020

Related benchmarks

TaskDatasetResultRank
Query-based meeting summarizationQMSum (test)
ROUGE-132.29
26
Meeting SummarizationICSI manual transcriptions--
22
Meeting SummarizationAMI
ROUGE-153.02
10
Meeting SummarizationICSI
ROUGE-146.28
9
Dialogue SummarizationAMI (test)
Conciseness4.05
9
Meeting SummarizationQMSum (all turns)
ROUGE-132.29
4
Meeting SummarizationAMI (test)
ROUGE-153.02
4
Meeting SummarizationQMSum Gold turns only
ROUGE-136.06
3
SummarizationAMI original (test)
BERTScore82.24
3
Dialogue SummarizationICSI (test)
Readability3.21
2
Showing 10 of 10 rows

Other info

Follow for update