Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

SliceFine: The Universal Winning-Slice Hypothesis for Pretrained Networks

About

This paper presents a theoretical framework explaining why fine tuning small, randomly selected subnetworks (slices) within pre trained models can be sufficient for downstream adaptation. We prove that pretrained networks exhibit a universal winning slice property arising from two phenomena: (1) spectral balance the eigenspectra of different weight matrix slices are remarkably similar; and (2) high task energy their backbone representations retain rich, task relevant features. This leads to the Universal Winning Slice Hypothesis, which provides a theoretical foundation for parameter efficient fine tuning (PEFT) in large scale models. Inspired by this, we propose SliceFine, a PEFT method that exploits this inherent redundancy by updating only selected slices of the original weights introducing zero new parameters, unlike adapter-based approaches. Empirically, SliceFine matches the performance of state of the art PEFT methods across language and vision tasks, while significantly improving training speed, memory efficiency, and model compactness. Our work bridges theory and practice, offering a theoretically grounded alternative to existing PEFT techniques.

Md Kowsher, Ali O. Polat, Ehsan Mohammady Ardehaly, Mehrdad Salehi, Zia Ghiasi, Prasanth Murali, Chen Chen• 2025

Related benchmarks

TaskDatasetResultRank
Image ClassificationMMT-47
Accuracy93.98
17
Action UnderstandingMMT-47 Action Understanding
Accuracy51.06
17
Commonsense ReasoningMMT-47 Commonsense Reasoning
Accuracy83.84
17
High-Level ReasoningMMT-47 High Level Reasoning
Accuracy43.3
17
Object Motion and Spatial ReasoningMMT-47 Object Motion & Spatial
Accuracy62.91
17
Natural Language UnderstandingGLUE
Accuracy90.54
17
Vision UnderstandingMMT-47 Vision Benchmark
Accuracy77.06
17
Showing 7 of 7 rows

Other info

Follow for update