Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Pretraining A Large Language Model using Distributed GPUs: A Memory-Efficient Decentralized Paradigm

About

Pretraining large language models (LLMs) typically requires centralized clusters with thousands of high-memory GPUs (e.g., H100/A100). Recent decentralized training methods reduce communication overhead by employing federated optimization; however, they still need to train the entire model on each node, remaining constrained by GPU memory limitations. In this work, we propose SParse Expert Synchronization (SPES), a memory-efficient decentralized framework for pretraining mixture-of-experts (MoE) LLMs. SPES trains only a subset of experts per node, substantially lowering the memory footprint. Each node updates its local experts and periodically synchronizes with other nodes, eliminating full-parameter transmission while ensuring efficient knowledge sharing. To accelerate convergence, we introduce an expert-merging warm-up strategy, where experts exchange knowledge early in training, to rapidly establish foundational capabilities. With SPES, we train a 2B-parameter MoE LLM using 16 standalone 48GB GPUs over internet connections, which achieves competitive performance with centrally trained LLMs under similar computational budgets. We further demonstrate scalability by training a 7B model from scratch and a 9B model upcycled from a dense checkpoint, both of which match prior centralized baselines. Our code is available at https://github.com/zjr2000/SPES.

Jinrui Zhang, Chaodong Xiao, Aoqi Wu, Xindong Zhang, Lei Zhang• 2026

Related benchmarks

TaskDatasetResultRank
Multi-task Language UnderstandingMMLU
Accuracy63.7
842
Question AnsweringARC Challenge
Accuracy57.3
749
Commonsense ReasoningPIQA
Accuracy78.9
647
Question AnsweringARC Easy
Normalized Acc81.5
385
Boolean Question AnsweringBoolQ
Accuracy77.3
307
Question AnsweringOBQA
Accuracy42.2
276
Question AnsweringSciQ
Accuracy95.3
226
Commonsense ReasoningSIQA
Accuracy47.5
96
Logical reasoningLogiQA
Accuracy30.4
84
Multi-level multi-discipline evaluationC-Eval
Accuracy44.7
28
Showing 10 of 10 rows

Other info

GitHub

Follow for update