Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed Mixture-of-Experts

About

Weight-sharing supernets are crucial for performance estimation in cutting-edge neural architecture search (NAS) frameworks. Despite their ability to generate diverse subnetworks without retraining, the quality of these subnetworks is not guaranteed due to weight sharing. In NLP tasks like machine translation and pre-trained language modeling, there is a significant performance gap between supernet and training from scratch for the same model architecture, necessitating retraining post optimal architecture identification. This study introduces a solution called mixture-of-supernets, a generalized supernet formulation leveraging mixture-of-experts (MoE) to enhance supernet model expressiveness with minimal training overhead. Unlike conventional supernets, this method employs an architecture-based routing mechanism, enabling indirect sharing of model weights among subnetworks. This customization of weights for specific architectures, learned through gradient descent, minimizes retraining time, significantly enhancing training efficiency in NLP. The proposed method attains state-of-the-art (SoTA) performance in NAS for fast machine translation models, exhibiting a superior latency-BLEU tradeoff compared to HAT, the SoTA NAS framework for machine translation. Furthermore, it excels in NAS for building memory-efficient task-agnostic BERT models, surpassing NAS-BERT and AutoDistil across various model sizes. The code can be found at: https://github.com/UBC-NLP/MoS.

Ganesh Jawahar, Haichuan Yang, Yunyang Xiong, Zechun Liu, Dilin Wang, Fei Sun, Meng Li, Aasish Pappu, Barlas Oguz, Muhammad Abdul-Mageed, Laks V. S. Lakshmanan, Raghuraman Krishnamoorthi, Vikas Chandra• 2023

Related benchmarks

TaskDatasetResultRank
Machine TranslationWMT En-Fr 2014 (test)
BLEU41.6
237
Performance PredictionWMT En-De 2019 (val)
MAE0.87
16
Performance PredictionWMT En-De 2014 (val)
MAE0.87
16
Performance PredictionWMT En-Fr 2014 (val)
MAE1.16
16
Performance PredictionWMT Benchmarks Average WMT'14 & WMT'19 (aggregation)
MAE0.97
16
Performance PredictionChatino to Spanish
MAE0.01
4
Performance PredictionBribri to Spanish
MAE0.02
4
Performance PredictionAmericasNLP Spanish to Bribri 2023 (test)
MAE0.43
4
Performance PredictionAmericasNLP Bribri to Spanish 2023 (test)
MAE0.29
4
Performance PredictionAmericasNLP Chatino to Spanish 2023 (test)
MAE2.42
4
Showing 10 of 10 rows

Other info

Follow for update