Partially Shared Concept Bottleneck Models
About
Concept Bottleneck Models (CBMs) enhance interpretability by introducing a layer of human-understandable concepts between inputs and predictions. While recent methods automate concept generation using Large Language Models (LLMs) and Vision-Language Models (VLMs), they still face three fundamental challenges: poor visual grounding, concept redundancy, and the absence of principled metrics to balance predictive accuracy and concept compactness. We introduce PS-CBM, a Partially Shared CBM framework that addresses these limitations through three core components: (1) a multimodal concept generator that integrates LLM-derived semantics with exemplar-based visual cues; (2) a Partially Shared Concept Strategy that merges concepts based on activation patterns to balance specificity and compactness; and (3) Concept-Efficient Accuracy (CEA), a post-hoc metric that jointly captures both predictive accuracy and concept compactness. Extensive experiments on eleven diverse datasets show that PS-CBM consistently outperforms state-of-the-art CBMs, improving classification accuracy by 1.0%-7.4% and CEA by 2.0%-9.5%, while requiring significantly fewer concepts. These results underscore PS-CBM's effectiveness in achieving both high accuracy and strong interpretability.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | Food-101 | Accuracy93 | 494 | |
| Image Classification | Flowers102 | Accuracy97.9 | 478 | |
| Image Classification | Food101 | Accuracy83 | 309 | |
| Image Classification | RESISC45 | -- | 263 | |
| Image Classification | CUB-200 2011 | Accuracy85.3 | 257 | |
| Image Classification | Oxford Flowers 102 | -- | 172 | |
| Image Classification | ImageNet | Acc84.5 | 45 | |
| Medical Image Classification | HAM10000 | Accuracy83.4 | 39 | |
| Image Classification | FGVC Aircraft | Accuracy65.1 | 32 | |
| Action Recognition | UCF-101 | Accuracy (ACC)90.4 | 10 |