Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Measuring Massive Multitask Language Understanding

About

We propose a new test to measure a text model's multitask accuracy. The test covers 57 tasks including elementary mathematics, US history, computer science, law, and more. To attain high accuracy on this test, models must possess extensive world knowledge and problem solving ability. We find that while most recent models have near random-chance accuracy, the very largest GPT-3 model improves over random chance by almost 20 percentage points on average. However, on every one of the 57 tasks, the best models still need substantial improvements before they can reach expert-level accuracy. Models also have lopsided performance and frequently do not know when they are wrong. Worse, they still have near-random accuracy on some socially important subjects such as morality and law. By comprehensively evaluating the breadth and depth of a model's academic and professional understanding, our test can be used to analyze models across many tasks and to identify important shortcomings.

Dan Hendrycks, Collin Burns, Steven Basart, Andy Zou, Mantas Mazeika, Dawn Song, Jacob Steinhardt• 2020

Related benchmarks

TaskDatasetResultRank
Commonsense ReasoningHellaSwag
Accuracy78.9
1460
Multi-task Language UnderstandingMMLU
Accuracy89.8
842
Commonsense ReasoningPIQA
Accuracy81
647
Multitask Language UnderstandingMMLU (test)
Accuracy89.8
303
Reading ComprehensionRACE high
Accuracy46.8
295
Multitask Language UnderstandingMMLU
Accuracy89.8
206
Reading ComprehensionRACE mid
Accuracy58.1
196
Common Sense ReasoningWinoGrande
Accuracy70.2
156
Common Sense ReasoningBoolQ
Accuracy60.5
131
Reading ComprehensionLAMBADA
Accuracy76.2
10
Showing 10 of 14 rows

Other info

Code

Follow for update