Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Trust the uncertain teacher: distilling dark knowledge via calibrated uncertainty

About

The core of knowledge distillation lies in transferring the teacher's rich 'dark knowledge'-subtle probabilistic patterns that reveal how classes are related and the distribution of uncertainties. While this idea is well established, teachers trained with conventional cross-entropy often fail to preserve such signals. Their distributions collapse into sharp, overconfident peaks that appear decisive but are in fact brittle, offering little beyond the hard label or subtly hindering representation-level transfer. This overconfidence is especially problematic in high-cardinality tasks, where the nuances among many plausible classes matter most for guiding a compact student. Moreover, such brittle targets reduce robustness under distribution shift, leaving students vulnerable to miscalibration in real-world conditions. To address this limitation, we revisit distillation from a distributional perspective and propose Calibrated Uncertainty Distillation (CUD), a framework designed to make dark knowledge more faithfully accessible. Instead of uncritically adopting the teacher's overconfidence, CUD encourages teachers to reveal uncertainty where it is informative and guides students to learn from targets that are calibrated rather than sharpened certainty. By directly shaping the teacher's predictive distribution before transfer, our approach balances accuracy and calibration, allowing students to benefit from both confident signals on easy cases and structured uncertainty on hard ones. Across diverse benchmarks, CUD yields students that are not only more accurate, but also more calibrated under shift and more reliable on ambiguous, long-tail inputs.

Jeonghyun Kim, SooKyung Kim, Richeng Xuan, Hyunsoo Cho• 2026

Related benchmarks

TaskDatasetResultRank
Intent ClassificationBanking77 (test)
Accuracy93.83
151
Question ClassificationTREC (test)
Accuracy97.4
124
Topic ClassificationAG News (test)
Accuracy94.91
98
Intent ClassificationClinc150 (test)
Accuracy96.07
26
Binary ClassificationGLUE (test)
QNLI Accuracy89.23
25
Out-of-Distribution DetectionCLINC Full (test)
AUROC94.52
21
Intent ClassificationMASSIVE (test)
In-Scope Accuracy89.44
17
Out-of-Distribution DetectionTREC (test)
FPR@908.6
8
CalibrationTREC--
8
CalibrationBanking77
ECE0.277
4
Showing 10 of 20 rows

Other info

Follow for update