GAPX: Generalized Autoregressive Paraphrase-Identification X
About
Paraphrase Identification is a fundamental task in Natural Language Processing. While much progress has been made in the field, the performance of many state-of-the-art models often suffer from distribution shift during inference time. We verify that a major source of this performance drop comes from biases introduced by negative examples. To overcome these biases, we propose in this paper to train two separate models, one that only utilizes the positive pairs and the other the negative pairs. This enables us the option of deciding how much to utilize the negative model, for which we introduce a perplexity based out-of-distribution metric that we show can effectively and automatically determine how much weight it should be given during inference. We support our findings with strong empirical results.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Paraphrase Detection | QQP (test) | Accuracy84.5 | 51 | |
| Paraphrase Identification | PAWS -> PAWS (test) | Accuracy92.7 | 22 | |
| Paraphrase Identification | PIT -> PAWS (test) | AUROC77.5 | 20 | |
| Paraphrase Identification | PAWS -> QQP (test) | AUROC77.7 | 20 | |
| Paraphrase Identification | QQP Out-of-distribution from PAWS | Macro F170.8 | 20 | |
| Paraphrase Identification | QQP -> WMT (test) | AUROC85.1 | 10 | |
| Paraphrase Identification | PIT -> WMT (test) | AUROC0.85 | 10 | |
| Paraphrase Identification | PAWS -> WMT (test) | AUROC0.849 | 10 | |
| Paraphrase Identification | WMT Out-of-distribution from QQP | Macro F175.5 | 10 | |
| Paraphrase Identification | QQP Out-of-distribution from PIT | Macro F10.757 | 10 |