Universal Language Model Fine-tuning for Text Classification
About
Inductive transfer learning has greatly impacted computer vision, but existing approaches in NLP still require task-specific modifications and training from scratch. We propose Universal Language Model Fine-tuning (ULMFiT), an effective transfer learning method that can be applied to any task in NLP, and introduce techniques that are key for fine-tuning a language model. Our method significantly outperforms the state-of-the-art on six text classification tasks, reducing the error by 18-24% on the majority of datasets. Furthermore, with only 100 labeled examples, it matches the performance of training from scratch on 100x more data. We open-source our pretrained models and code.
Jeremy Howard, Sebastian Ruder• 2018
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Mathematical Reasoning | GSM8K (test) | Accuracy63 | 797 | |
| Image Classification | DTD | Accuracy67.85 | 419 | |
| Image Classification | SVHN | Accuracy94 | 359 | |
| Image Classification | FGVCAircraft | Accuracy54.77 | 225 | |
| Text Classification | AG News (test) | Accuracy84.14 | 210 | |
| Text Classification | TREC | Accuracy96.4 | 179 | |
| Sentiment Classification | IMDB (test) | Error Rate4.6 | 144 | |
| Text Classification | Yahoo! Answers (test) | Clean Accuracy64.27 | 133 | |
| Image Classification | VTAB 1k (test) | -- | 121 | |
| Text Classification | AGNews | Accuracy95 | 119 |
Showing 10 of 59 rows