Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

FAMO: Fast Adaptive Multitask Optimization

About

One of the grand enduring goals of AI is to create generalist agents that can learn multiple different tasks from diverse data via multitask learning (MTL). However, in practice, applying gradient descent (GD) on the average loss across all tasks may yield poor multitask performance due to severe under-optimization of certain tasks. Previous approaches that manipulate task gradients for a more balanced loss decrease require storing and computing all task gradients ($\mathcal{O}(k)$ space and time where $k$ is the number of tasks), limiting their use in large-scale scenarios. In this work, we introduce Fast Adaptive Multitask Optimization FAMO, a dynamic weighting method that decreases task losses in a balanced way using $\mathcal{O}(1)$ space and time. We conduct an extensive set of experiments covering multi-task supervised and reinforcement learning problems. Our results indicate that FAMO achieves comparable or superior performance to state-of-the-art gradient manipulation techniques while offering significant improvements in space and computational efficiency. Code is available at \url{https://github.com/Cranial-XIX/FAMO}.

Bo Liu, Yihao Feng, Peter Stone, Qiang Liu• 2023

Related benchmarks

TaskDatasetResultRank
Semantic segmentationCityscapes (test)
mIoU74.54
1154
Semantic segmentationCityscapes
mIoU74.76
658
Depth EstimationNYU v2 (test)--
432
Surface Normal EstimationNYU v2 (test)
Mean Angle Distance (MAD)25.06
224
Semantic segmentationNYU Depth V2 (test)
mIoU38.88
183
Surface Normal PredictionNYU V2
Mean Error24.5
118
Depth EstimationCityscapes
Abs. Err.0.014
53
Multi-task LearningCityscapes (test)
MR5.5
43
Depth EstimationCityscapes (test)
Abs Err0.0145
40
Depth EstimationCityscapes
Absolute Error0.014
34
Showing 10 of 31 rows

Other info

Follow for update