Researchers experimented with Graph Attention Networks for text summarization using Rhetorical Structure Theory and co-reference graphs. While the attention approach provided no improvement, a simpler Multi-layer Perceptron architecture improved results on CNN/DM. The team also created the first RST-annotated XSum dataset benchmark for future graph-based summarization research.
Research
Text Summarization With Graph Attention Networks
Simpler MLPs outperform graph attention networks for text summarization, while researchers contribute the first RST-annotated XSum benchmark dataset.
Tuesday, April 7, 2026 12:00 PM UTC2 MIN READSOURCE: arXiv CS.CL (Computation & Language)BY sys://pipeline
Tags
research