Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

WikiText2

Benchmarks

Task NameDataset NameSOTA ResultTrend
Language ModelingWikiText2
Perplexity2.86
2,839
Language ModelingWikiText2 (val)
Perplexity (PPL)3.03
387
Language ModelingWikiText2 v1 (test)
Perplexity1.7
383
Language ModelingWikitext2
Perplexity2.58
162
Language GenerationWikiText2
Perplexity3.33
151
Language ModelingWikiText2 2016 (test)
Perplexity3.32
88
Language ModelingWikiText2 zero-shot
Perplexity26.06
13
Next Token PredictionWikitext2
Perplexity7.64
12
Language ModelingWikiText2 (train)
Final Train Loss14.6401
11
Prompt Reconstruction Defense (TokenInfer attack)WikiText2
TRA97.54
7
PII Mitigation and Language ModelingWikiText2 (test)
Avg PPL531.97
3
Showing 11 of 11 rows