It's Morphin' Time! Combating Linguistic Discrimination with Inflectional Perturbations
About
Training on only perfect Standard English corpora predisposes pre-trained neural networks to discriminate against minorities from non-standard linguistic backgrounds (e.g., African American Vernacular English, Colloquial Singapore English, etc.). We perturb the inflectional morphology of words to craft plausible and semantically similar adversarial examples that expose these biases in popular NLP models, e.g., BERT and Transformer, and show that adversarially fine-tuning them for a single epoch significantly improves robustness without sacrificing performance on clean data.
Samson Tan, Shafiq Joty, Min-Yen Kan, Richard Socher• 2020
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Text Translation | En-Fr | BLEU Score0.37 | 14 | |
| Text Translation | En-Zh | BLEU0.3 | 14 | |
| Speech Translation | CoVoST2 En-De | BLEU18.2 | 10 | |
| Speech-to-text Translation | CoVoST2 fr-en | BLEU24.7 | 8 | |
| Speech Translation | CoVoST 2 En-De Morpheus adversarial | BLEU19.9 | 7 | |
| Speech Translation | CoVoST 2 En-Ca (Morpheus adversarial) | BLEU Score26.1 | 7 | |
| Speech Translation | CoVoST 2 En-Ar Morpheus adversarial | BLEU15.4 | 7 | |
| Speech Translation | CoVoST 2 Fr-En Morpheus adversarial | BLEU25.1 | 7 | |
| Speech Translation | CoVoST 2 En-De (Clean TTS) | BLEU22.5 | 7 | |
| Speech Translation | CoVoST 2 En-Ca Clean TTS | BLEU28.7 | 7 |
Showing 10 of 18 rows