Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

ROCKET: Rapid Optimization via Calibration-guided Knapsack Enhanced Truncation for Efficient Model Compression

About

We present ROCKET, a training-free model compression method that achieves state-of-the-art performance in comparison with factorization, structured-sparsification and dynamic compression baselines. Operating under a global compression budget, ROCKET comprises two key innovations: First, it formulates layer-wise compression allocation as a multi-choice knapsack problem, selecting the optimal compression level for each layer to minimize total reconstruction error while adhering to a target model size. Second, it introduces a single-step sparse matrix factorization inspired by dictionary learning: using only a small calibration set, it sparsifies weight coefficients based on activation-weights sensitivity and then updates the dictionary in closed form via least squares bypassing iterative optimization, sparse coding, or backpropagation entirely. ROCKET consistently outperforms existing compression approaches across different model architectures at 20-50\% compression rates. Notably, it retains over 90\% of the original model's performance at 30\% compression without any fine-tuning. Moreover, when applying a light fine-tuning phase, recovery is substantially enhanced: for instance, compressing Qwen3-14B to an 8B-parameter model and healing it with just 30 million tokens yields performance nearly on par with the original Qwen3-8B. The code for ROCKET is at github.com/mts-ai/ROCKET/tree/main.

Ammar Ali, Baher Mohammad, Denis Makhov, Dmitriy Shopkhoev, Magauiya Zhussip, Stamatios Lefkimmiatis• 2026

Related benchmarks

TaskDatasetResultRank
Commonsense ReasoningHellaSwag
Accuracy54
1891
Commonsense ReasoningWinoGrande
Accuracy72
1085
Question AnsweringARC Challenge
Accuracy42
906
Commonsense ReasoningPIQA
Accuracy78
751
Instruction FollowingIFEval--
625
Question AnsweringARC Easy
Accuracy76
597
Mathematical Problem SolvingMATH
Accuracy11.1
229
Language ModelingWikiText and LAMBADA
WikiText Perplexity15
47
Complex ReasoningBBH
Accuracy50.33
40
Zero-shot ClassificationAccuracy Benchmarks (PIQA, HellaSwag, LAMBADA, ARC-e, ARC-c, SciQ, Race, MMLU) Zero-shot
PIQA77.6
39
Showing 10 of 18 rows

Other info

GitHub

Follow for update