Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Text Summarization With Graph Attention Networks

About

This study aimed to leverage graph information, particularly Rhetorical Structure Theory (RST) and Co-reference (Coref) graphs, to enhance the performance of our baseline summarization models. Specifically, we experimented with a Graph Attention Network architecture to incorporate graph information. However, this architecture did not enhance the performance. Subsequently, we used a simple Multi-layer Perceptron architecture, which improved the results in our proposed model on our primary dataset, CNN/DM. Additionally, we annotated XSum dataset with RST graph information, establishing a benchmark for future graph-based summarization models. This secondary dataset posed multiple challenges, revealing both the merits and limitations of our models.

Mohammadreza Ardestani, Yllias Chali• 2026

Related benchmarks

TaskDatasetResultRank
SummarizationXSum (test)
ROUGE-214.66
246
SummarizationCNNDM (test)
ROUGE 220.91
31
Showing 2 of 2 rows

Other info

Follow for update