| Task Name | Dataset Name | SOTA Result | Trend | |
|---|---|---|---|---|
| Language Modeling | WikiText2 | Perplexity2.86 | 1,875 | |
| Language Modeling | WikiText2 v1 (test) | Perplexity1.7 | 341 | |
| Language Modeling | WikiText2 (val) | Perplexity (PPL)3.03 | 277 | |
| Language Modeling | WikiText2 2016 (test) | Perplexity3.32 | 88 | |
| Language Generation | WikiText2 | Perplexity3.53 | 36 | |
| Language Modeling | WikiText2 zero-shot | Perplexity26.06 | 13 | |
| Next Token Prediction | Wikitext2 | Perplexity7.64 | 12 | |
| Language Modeling | WikiText2 (train) | Final Train Loss14.6401 | 11 | |
| PII Mitigation and Language Modeling | WikiText2 (test) | Avg PPL531.97 | 3 |