Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

FiGKD: Fine-Grained Knowledge Distillation via High-Frequency Detail Transfer

About

Knowledge distillation (KD) is a widely adopted technique for transferring knowledge from a high-capacity teacher model to a smaller student model by aligning their output distributions. However, existing methods often underperform in fine-grained visual recognition tasks, where distinguishing subtle differences between visually similar classes is essential. This performance gap stems from the fact that conventional approaches treat the teacher's output logits as a single, undifferentiated signal-assuming all contained information is equally beneficial to the student. Consequently, student models may become overloaded with redundant signals and fail to capture the teacher's nuanced decision boundaries. To address this issue, we propose Fine-Grained Knowledge Distillation (FiGKD), a novel frequency-aware framework that decomposes a model's logits into low-frequency (content) and high-frequency (detail) components using the discrete wavelet transform (DWT). FiGKD selectively transfers only the high-frequency components, which encode the teacher's semantic decision patterns, while discarding redundant low-frequency content already conveyed through ground-truth supervision. Our approach is simple, architecture-agnostic, and requires no access to intermediate feature maps. Extensive experiments on CIFAR-100, TinyImageNet, and multiple fine-grained recognition benchmarks show that FiGKD consistently outperforms state-of-the-art logit-based and feature-based distillation methods across a variety of teacher-student configurations. These findings confirm that frequency-aware logit decomposition enables more efficient and effective knowledge transfer, particularly in resource-constrained settings.

Seonghak Kim• 2025

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (val)--
776
Image ClassificationTinyImageNet (val)--
289
Image ClassificationCUB-200--
106
Image ClassificationDogs
Accuracy70.14
68
Image ClassificationStanford 40 Actions (test)
Top-1 Accuracy48.99
48
Fine-grained Visual RecognitionCUB200 (val)
Top-1 Accuracy71.97
24
Fine-grained Visual RecognitionStanford40 (val)
Top-1 Accuracy60.21
24
Fine-grained Visual RecognitionStanford Dogs (val)
Top-1 Acc74.38
24
Scene recognitionMIT indoor 67 (val)
Top-1 Accuracy68.43
24
Showing 9 of 9 rows

Other info

Follow for update