Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

LiME: Lightweight Mixture of Experts for Efficient Multimodal Multi-task Learning

About

MoE-PEFT methods combine Mixture of Experts with parameter-efficient fine-tuning for multi-task adaptation, but require separate adapters per expert causing trainable parameters to scale linearly with expert count and limiting applicability to adapter-based architectures. We propose LiME (Lightweight Mixture of Experts), which achieves expert specialization through lightweight modulation rather than adapter replication. Instead of separate adapters, LiME uses a single shared PEFT module and modulates its output with lightweight expert vectors, reducing expert parameters while generalizing to any PEFT method. Notably, LiME introduces zero-parameter routing by leveraging existing frozen and adapted representations eliminating learned router parameters typically required per layer. Theoretically, we prove that (i) more experts preserve more task-relevant information and (ii) modulation approximates full expert-specific PEFT with bounded error. LiME further incorporates n-gram windowed routing and adaptive expert selection (Auto Top-K) based on routing confidence. Experiments on MMT-47, a multimodal multi-task benchmark with 47 tasks spanning text, image, and video, demonstrate that LiME achieves competitive or superior performance while using up to 4x fewer trainable parameters and up to 29% faster training compared to corresponding MoE-PEFT baselines.

Md Kowsher, Haris Mansoor, Nusrat Jahan Prottasha, Ozlem Garibay, Victor Zhu, Zhengping Ji, Chen Chen• 2026

Related benchmarks

TaskDatasetResultRank
Commonsense ReasoningMMT-47 Commonsense Reasoning
Accuracy84.98
17
Object Motion and Spatial ReasoningMMT-47 Object Motion & Spatial
Accuracy65.41
17
Vision UnderstandingMMT-47 Vision Benchmark
Accuracy78.12
17
Action UnderstandingMMT-47 Action Understanding
Accuracy53.39
17
Natural Language UnderstandingGLUE
Accuracy91.19
17
High-Level ReasoningMMT-47 High Level Reasoning
Accuracy45.65
17
Image ClassificationMMT-47
Accuracy94.61
17
Language UnderstandingGLUE (test)
SST-2 Score95.99
6
Showing 8 of 8 rows

Other info

Follow for update