| Task Name | Dataset Name | SOTA Result | Trend | |
|---|---|---|---|---|
| Language Modeling | WikiText2 | Perplexity2.86 | 2,839 | |
| Language Modeling | WikiText2 (val) | Perplexity (PPL)3.03 | 387 | |
| Language Modeling | WikiText2 v1 (test) | Perplexity1.7 | 383 | |
| Language Modeling | Wikitext2 | Perplexity2.58 | 162 | |
| Language Generation | WikiText2 | Perplexity3.33 | 151 | |
| Language Modeling | WikiText2 2016 (test) | Perplexity3.32 | 88 | |
| Language Modeling | WikiText2 zero-shot | Perplexity26.06 | 13 | |
| Next Token Prediction | Wikitext2 | Perplexity7.64 | 12 | |
| Language Modeling | WikiText2 (train) | Final Train Loss14.6401 | 11 | |
| Prompt Reconstruction Defense (TokenInfer attack) | WikiText2 | TRA97.54 | 7 | |
| PII Mitigation and Language Modeling | WikiText2 (test) | Avg PPL531.97 | 3 |