Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Promoting Graph Awareness in Linearized Graph-to-Text Generation

About

Generating text from structured inputs, such as meaning representations or RDF triples, has often involved the use of specialized graph-encoding neural networks. However, recent applications of pretrained transformers to linearizations of graph inputs have yielded state-of-the-art generation results on graph-to-text tasks. Here, we explore the ability of these linearized models to encode local graph structures, in particular their invariance to the graph linearization strategy and their ability to reconstruct corrupted inputs. Our findings motivate solutions to enrich the quality of models' implicit graph encodings via scaffolding. Namely, we use graph-denoising objectives implemented in a multi-task text-to-text framework. We find that these denoising scaffolds lead to substantial improvements in downstream generation in low-resource settings.

Alexander Hoyle, Ana Marasovi\'c, Noah Smith• 2020

Related benchmarks

TaskDatasetResultRank
AMR-to-text generationLDC2017T10 (test)
BLEU44.9
55
Showing 1 of 1 rows

Other info

Follow for update