Adversarial Self-Attention for Language Understanding
About
Deep neural models (e.g. Transformer) naturally learn spurious features, which create a ``shortcut'' between the labels and inputs, thus impairing the generalization and robustness. This paper advances the self-attention mechanism to its robust variant for Transformer-based pre-trained language models (e.g. BERT). We propose \textit{Adversarial Self-Attention} mechanism (ASA), which adversarially biases the attentions to effectively suppress the model reliance on features (e.g. specific keywords) and encourage its exploration of broader semantics. We conduct a comprehensive evaluation across a wide range of tasks for both pre-training and fine-tuning stages. For pre-training, ASA unfolds remarkable performance gains compared to naive training for longer steps. For fine-tuning, ASA-empowered models outweigh naive models by a large margin considering both generalization and robustness.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Natural Language Understanding | GLUE | SST-296.3 | 452 | |
| Named Entity Recognition | Wnut 2017 | F1 Score57.3 | 79 | |
| Paraphrase Detection | PAWS QQP | Accuracy96 | 16 | |
| Dialogue Comprehension | DREAM | Accuracy69.2 | 15 | |
| Common Sense Reasoning | HELLASWAG (dev) | Accuracy95.4 | 12 | |
| Natural Language Inference | ANLI all rounds (test) | Accuracy58.2 | 4 |