| Task Name | Dataset Name | SOTA Result | Trend | |
|---|---|---|---|---|
| Natural Language Understanding | GLUE (dev) | SST-2 (Acc)97.36 | 504 | |
| Natural Language Understanding | GLUE | SST-2156 | 452 | |
| Natural Language Understanding | GLUE (test) | SST-2 Accuracy97.9 | 416 | |
| Natural Language Understanding | GLUE (val) | SST-297.4 | 170 | |
| Natural Language Understanding | GLUE (test dev) | MRPC Accuracy93.45 | 81 | |
| General Language Understanding | GLUE | Accuracy92.5 | 66 | |
| Natural Language Understanding | GLUE (test val) | MRPC Accuracy94 | 59 | |
| Natural Language Understanding | GLUE | COLA Score69.3 | 41 | |
| General Language Understanding | GLUE v1 (test dev) | MNLI87.86 | 40 | |
| Natural Language Understanding | GLUE (test) | SST-2 Accuracy95.64 | 33 | |
| Adversarial Attack | GLUE | SST-2 Speedup3.56 | 32 | |
| Natural Language Understanding | GLUE | SST-2 Speedup3.06 | 32 | |
| Natural Language Understanding | GLUE v1 (dev) | MRPC Score93.8 | 30 | |
| Natural Language Understanding | GLUE | SST-2 Acc97.5 | 28 | |
| Natural Language Understanding | GLUE (test) | QNLI Score92.31 | 26 | |
| Natural Language Understanding | GLUE (test) | MNLI-mm98.6 | 26 | |
| Binary classification | GLUE (test) | QNLI Accuracy90.66 | 25 | |
| Natural Language Understanding | GLUE 1.0 (test) | CoLA (MCC)66.4 | 25 | |
| Natural Language Understanding | GLUE RoBERTa LARGE (test dev) | MNLI Accuracy90.57 | 22 | |
| Natural Language Inference | GLUE (test) | MNLI Acc93.15 | 18 | |
| Natural Language Understanding | GLUE official (val test) | SST-2 Accuracy0.97 | 18 | |
| Natural Language Understanding | GLUE RoBERTa-base (val) | CoLA Score60.18 | 16 | |
| Natural Language Understanding | GLUE | CoLA76.98 | 16 | |
| Natural Language Understanding | GLUE | CoLA56.7 | 16 | |
| Natural Language Understanding | GLUE | CoLA Score63.1 | 15 |