Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ForkMerge: Mitigating Negative Transfer in Auxiliary-Task Learning

About

Auxiliary-Task Learning (ATL) aims to improve the performance of the target task by leveraging the knowledge obtained from related tasks. Occasionally, learning multiple tasks simultaneously results in lower accuracy than learning only the target task, which is known as negative transfer. This problem is often attributed to the gradient conflicts among tasks, and is frequently tackled by coordinating the task gradients in previous works. However, these optimization-based methods largely overlook the auxiliary-target generalization capability. To better understand the root cause of negative transfer, we experimentally investigate it from both optimization and generalization perspectives. Based on our findings, we introduce ForkMerge, a novel approach that periodically forks the model into multiple branches, automatically searches the varying task weights by minimizing target validation errors, and dynamically merges all branches to filter out detrimental task-parameter updates. On a series of auxiliary-task learning benchmarks, ForkMerge outperforms existing methods and effectively mitigates negative transfer.

Junguang Jiang, Baixu Chen, Junwei Pan, Ximei Wang, Liu Dapeng, Jie Jiang, Mingsheng Long• 2023

Related benchmarks

TaskDatasetResultRank
Depth EstimationNYU v2 (test)--
423
Semantic segmentationNYU v2 (test)
mIoU53.67
248
Surface Normal EstimationNYU v2 (test)
Mean Angle Distance (MAD)22.18
206
Multi-task LearningNYU v2 (test)
Delta m%403
31
Image RecognitionDomainNet 50% test split (val)
Accuracy (Clipart)79.9
16
Multi-task RecommendationAliExpress (test)
CTR ES0.7402
16
Semi-supervised Image ClassificationCIFAR-10 4000 labeled samples (train test)
Test Error13.1
14
Semi-Supervised LearningSVHN 1000 labels (test)
Test Error5.49
14
Showing 8 of 8 rows

Other info

Code

Follow for update