A Joint Many-Task Model: Growing a Neural Network for Multiple NLP Tasks
About
Transfer and multi-task learning have traditionally focused on either a single source-target pair or very few, similar tasks. Ideally, the linguistic levels of morphology, syntax and semantics would benefit each other by being trained in a single model. We introduce a joint many-task model together with a strategy for successively growing its depth to solve increasingly complex tasks. Higher layers include shortcut connections to lower-level task predictions to reflect linguistic hierarchies. We use a simple regularization term to allow for optimizing all model weights to improve one task's loss without exhibiting catastrophic interference of the other tasks. Our single end-to-end model obtains state-of-the-art or competitive results on five different tasks from tagging, parsing, relatedness, and entailment tasks.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Chunking | CoNLL 2000 (test) | F1 Score95.77 | 88 | |
| Dependency Parsing | WSJ (test) | UAS94.67 | 67 | |
| Part-of-Speech Tagging | Penn Treebank (test) | Accuracy97.55 | 64 | |
| Part-of-Speech Tagging | WSJ (test) | Accuracy97.55 | 51 | |
| Part-of-Speech Tagging | POS (test) | Accuracy97.55 | 33 | |
| Chunking | Chunk (test) | F1 Score95.77 | 28 | |
| Dependency Parsing | Dep. Parse (test) | UAS94.7 | 23 | |
| Textual Entailment | SICK (test) | Accuracy86.8 | 21 | |
| Dependency Parsing | Penn Treebank (PTB) Section 23 v2.2 (test) | UAS94.67 | 17 | |
| Parsing | English PTB-SD 3.3.0 (test) | UAS94.67 | 7 |