Reducing Sensitivity on Speaker Names for Text Generation from Dialogues
About
Changing speaker names consistently throughout a dialogue should not affect its meaning and corresponding outputs for text generation from dialogues. However, pre-trained language models, serving as the backbone for dialogue-processing tasks, have shown to be sensitive to nuances. This may result in unfairness in real-world applications. No comprehensive analysis of this problem has been done in the past. In this work, we propose to quantitatively measure a model's sensitivity on speaker names, and comprehensively evaluate a number of known methods for reducing speaker name sensitivity, including a novel approach of our own. Extensive experiments on multiple datasets provide a benchmark for this problem and show the favorable performance of our approach in sensitivity reduction and quality of generation.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Dialogue Summarization | SamSum (test) | ROUGE-228.73 | 80 | |
| Abstractive dialogue summarization | SamSum (test) | -- | 53 | |
| Question Generation | SQuAD 1.1 (test) | BLEU-419.71 | 29 | |
| Question Generation | Molweni (test) | BLEU Score20.26 | 8 | |
| Dialogue Summarization | SAMSum In-distribution Names (test) | R228.79 | 4 | |
| Dialogue Summarization | SAMSum All-possible Names (test) | R228.44 | 4 | |
| Question Generation | Molweni In-distribution Names | BLEU19.58 | 4 | |
| Question Generation | Molweni All-possible Names | BLEU19.57 | 4 | |
| Reading Comprehension | SQuAD 2.0 (test) | BLEU29.03 | 4 | |
| Reading Comprehension | Molweni (test) | BLEU29.44 | 4 |