Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

LumiNet: Perception-Driven Knowledge Distillation via Statistical Logit Calibration

About

In the knowledge distillation literature, feature-based methods have dominated due to their ability to effectively tap into extensive teacher models. In contrast, logit-based approaches, which aim to distill "dark knowledge" from teachers, typically exhibit inferior performance compared to feature-based methods. To bridge this gap, we present LumiNet, a novel knowledge distillation algorithm designed to enhance logit-based distillation. We introduce the concept of "perception", aiming to calibrate logits based on the model's representation capability. This concept addresses overconfidence issues in the logit-based distillation method while also introducing a novel method to distill knowledge from the teacher. It reconstructs the logits of a sample/instances by considering relationships with other samples in the batch. LumiNet excels on benchmarks like CIFAR-100, ImageNet, and MSCOCO, outperforming the leading feature-based methods, e.g., compared to KD with ResNet18 and MobileNetV2 on ImageNet, it shows improvements of 1.5% and 2.05%, respectively. Codes are available at https://github.com/ismail31416/LumiNet.

Md. Ismail Hossain, M M Lutfe Elahi, Sameera Ramasinghe, Ali Cheraghian, Fuad Rahman, Nabeel Mohammed, Shafin Rahman• 2023

Related benchmarks

TaskDatasetResultRank
Object DetectionCOCO 2017 (val)
AP35.34
2454
Image ClassificationImageNet (val)
Top-1 Acc72.55
1206
Image ClassificationCIFAR-100 (val)
Accuracy78.95
661
Image ClassificationCIFAR-100
Top-1 Accuracy79.94
622
Natural Language UnderstandingGLUE (dev)
SST-2 (Acc)91.3
504
Instruction FollowingSelfInst
R-L Score14.7
50
Instruction FollowingDolly
Score28.6
18
Instruction FollowingVicuna
Score17.5
18
Showing 8 of 8 rows

Other info

Code

Follow for update