Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed Mixture-of-Experts
About
Weight-sharing supernets are crucial for performance estimation in cutting-edge neural architecture search (NAS) frameworks. Despite their ability to generate diverse subnetworks without retraining, the quality of these subnetworks is not guaranteed due to weight sharing. In NLP tasks like machine translation and pre-trained language modeling, there is a significant performance gap between supernet and training from scratch for the same model architecture, necessitating retraining post optimal architecture identification. This study introduces a solution called mixture-of-supernets, a generalized supernet formulation leveraging mixture-of-experts (MoE) to enhance supernet model expressiveness with minimal training overhead. Unlike conventional supernets, this method employs an architecture-based routing mechanism, enabling indirect sharing of model weights among subnetworks. This customization of weights for specific architectures, learned through gradient descent, minimizes retraining time, significantly enhancing training efficiency in NLP. The proposed method attains state-of-the-art (SoTA) performance in NAS for fast machine translation models, exhibiting a superior latency-BLEU tradeoff compared to HAT, the SoTA NAS framework for machine translation. Furthermore, it excels in NAS for building memory-efficient task-agnostic BERT models, surpassing NAS-BERT and AutoDistil across various model sizes. The code can be found at: https://github.com/UBC-NLP/MoS.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Machine Translation | WMT En-Fr 2014 (test) | BLEU41.6 | 237 | |
| Performance Prediction | WMT En-De 2019 (val) | MAE0.87 | 16 | |
| Performance Prediction | WMT En-De 2014 (val) | MAE0.87 | 16 | |
| Performance Prediction | WMT En-Fr 2014 (val) | MAE1.16 | 16 | |
| Performance Prediction | WMT Benchmarks Average WMT'14 & WMT'19 (aggregation) | MAE0.97 | 16 | |
| Performance Prediction | Chatino to Spanish | MAE0.01 | 4 | |
| Performance Prediction | Bribri to Spanish | MAE0.02 | 4 | |
| Performance Prediction | AmericasNLP Spanish to Bribri 2023 (test) | MAE0.43 | 4 | |
| Performance Prediction | AmericasNLP Bribri to Spanish 2023 (test) | MAE0.29 | 4 | |
| Performance Prediction | AmericasNLP Chatino to Spanish 2023 (test) | MAE2.42 | 4 |