Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases

About

This paper addresses the growing need for efficient large language models (LLMs) on mobile devices, driven by increasing cloud costs and latency concerns. We focus on designing top-quality LLMs with fewer than a billion parameters, a practical choice for mobile deployment. Contrary to prevailing belief emphasizing the pivotal role of data and parameter quantity in determining model quality, our investigation underscores the significance of model architecture for sub-billion scale LLMs. Leveraging deep and thin architectures, coupled with embedding sharing and grouped-query attention mechanisms, we establish a strong baseline network denoted as MobileLLM, which attains a remarkable 2.7%/4.3% accuracy boost over preceding 125M/350M state-of-the-art models. Additionally, we propose an immediate block-wise weight-sharing approach with no increase in model size and only marginal latency overhead. The resultant models, denoted as MobileLLM-LS, demonstrate a further accuracy enhancement of 0.7%/0.8% than MobileLLM 125M/350M. Moreover, MobileLLM model family shows significant improvements compared to previous sub-billion models on chat benchmarks, and demonstrates close correctness to LLaMA-v2 7B in API calling tasks, highlighting the capability of small models for common on-device use cases.

Zechun Liu, Changsheng Zhao, Forrest Iandola, Chen Lai, Yuandong Tian, Igor Fedorov, Yunyang Xiong, Ernie Chang, Yangyang Shi, Raghuraman Krishnamoorthi, Liangzhen Lai, Vikas Chandra• 2024

Related benchmarks

TaskDatasetResultRank
Commonsense ReasoningHellaSwag
Accuracy38.7
1891
Question AnsweringARC Challenge--
906
Commonsense ReasoningPIQA
Accuracy68.1
751
Question AnsweringARC Easy
Accuracy41
597
Question AnsweringSciQ--
283
Language ModelingLAMBADA
Accuracy34.1
268
Reading ComprehensionRACE
Accuracy28.7
151
Multi-task Language UnderstandingMMLU
Accuracy24.1
111
Language ModelingWikiText (val)
Perplexity32.27
54
Commonsense ReasoningZero-shot Evaluation Suite (HellaSwag, PIQA, Arc-E, Arc-C, WinoGrande, OBQA, SIQA, BoolQ)
HellaSwag (Zero-shot)37.57
15
Showing 10 of 10 rows

Other info

Follow for update