Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Latent Multi-task Architecture Learning

About

Multi-task learning (MTL) allows deep neural networks to learn from related tasks by sharing parameters with other networks. In practice, however, MTL involves searching an enormous space of possible parameter sharing architectures to find (a) the layers or subspaces that benefit from sharing, (b) the appropriate amount of sharing, and (c) the appropriate relative weights of the different task losses. Recent work has addressed each of the above problems in isolation. In this work we present an approach that learns a latent multi-task architecture that jointly addresses (a)--(c). We present experiments on synthetic data and data from OntoNotes 5.0, including four different tasks and seven different domains. Our extension consistently outperforms previous approaches to learning latent architectures for multi-task problems and achieves up to 15% average error reductions over common approaches to MTL.

Sebastian Ruder, Joachim Bingel, Isabelle Augenstein, Anders S{\o}gaard• 2017

Related benchmarks

TaskDatasetResultRank
Semantic segmentationCityscapes (test)
mIoU39.8
1145
Depth EstimationNYU v2 (test)--
423
Semantic segmentationNYU v2 (test)
mIoU40.8
248
Surface Normal EstimationNYU v2 (test)
Mean Angle Distance (MAD)15.3
206
Depth EstimationNYU Depth V2--
177
Surface Normal PredictionNYU V2
Mean Error14.2
100
Semantic segmentationNYU V2
mIoU23.8
74
Multi-Task AdaptationPascal Context (test)--
70
Monocular Depth EstimationCityscapes
Accuracy (delta < 1.25)68.9
62
Multi-task LearningCityscapes (test)--
43
Showing 10 of 44 rows

Other info

Follow for update