Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Controlled Hallucinations: Learning to Generate Faithfully from Noisy Data

About

Neural text generation (data- or text-to-text) demonstrates remarkable performance when training data is abundant which for many applications is not the case. To collect a large corpus of parallel data, heuristic rules are often used but they inevitably let noise into the data, such as phrases in the output which cannot be explained by the input. Consequently, models pick up on the noise and may hallucinate--generate fluent but unsupported text. Our contribution is a simple but powerful technique to treat such hallucinations as a controllable aspect of the generated text, without dismissing any input and without modifying the model architecture. On the WikiBio corpus (Lebret et al., 2016), a particularly noisy dataset, we demonstrate the efficacy of the technique both in an automatic and in a human evaluation.

Katja Filippova• 2020

Related benchmarks

TaskDatasetResultRank
Data-to-text generationToTTo
BLEU19.48
18
Data-to-text generationWIKIBIO (test)
BLEU36.5
17
Table-to-text generationWikiBio (val)
Fluency93.5
4
Entity-level factuality evaluationMENT Dataset converted
Accuracy68.22
3
Factuality classificationXENT (test)
Accuracy84.54
3
Hallucination classificationXENT (test)
Accuracy92.93
3
Showing 6 of 6 rows

Other info

Follow for update