Shaping Parameter Contribution Patterns for Out-of-Distribution Detection
About
Out-of-distribution (OOD) detection is a well-known challenge due to deep models often producing overconfident. In this paper, we reveal a key insight that trained classifiers tend to rely on sparse parameter contribution patterns, meaning that only a few dominant parameters drive predictions. This brittleness can be exploited by OOD inputs that anomalously trigger these parameters, resulting in overconfident predictions. To address this issue, we propose a simple yet effective method called Shaping Parameter Contribution Patterns (SPCP), which enhances OOD detection robustness by encouraging the classifier to learn boundary-oriented dense contribution patterns. Specifically, SPCP operates during training by rectifying excessively high parameter contributions based on a dynamically estimated threshold. This mechanism promotes the classifier to rely on a broader set of parameters for decision-making, thereby reducing the risk of overconfident predictions caused by anomalously triggered parameters, while preserving in-distribution (ID) performance. Extensive experiments under various OOD detection setups verify the effectiveness of SPCP.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Out-of-Distribution Detection | ImageNet Far-OOD | AUROC94.11 | 52 | |
| Out-of-Distribution Detection | ImageNet-1K Near-OOD OpenOOD v1.5 | AUROC83.2 | 51 | |
| Out-of-Distribution Detection | CIFAR-100 | Near-OOD FPR9555.21 | 12 | |
| Out-of-Distribution Detection | CIFAR-10 | Near-OOD FPR9531.67 | 12 |