Towards Achieving Concept Completeness for Textual Concept Bottleneck Models
About
Textual Concept Bottleneck Models (TCBMs) are interpretable-by-design models for text classification that predict a set of salient concepts before making the final prediction. This paper proposes Complete Textual Concept Bottleneck Model (CT-CBM), a novel TCBM generator building concept labels in a fully unsupervised manner using a small language model, eliminating both the need for predefined human labeled concepts and LLM annotations. CT-CBM iteratively targets and adds important and identifiable concepts in the bottleneck layer to create a complete concept basis. CT-CBM achieves striking results against competitors in terms of concept basis completeness and concept detection accuracy, offering a promising solution to reliably enhance interpretability of NLP classifiers.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | CUB-200 (test) | Accuracy77.8 | 98 | |
| Text Classification | DBPedia (test) | -- | 40 | |
| News Classification | AG News (test) | Accuracy91.2 | 34 | |
| Classification | N24 News (test) | Accuracy98.5 | 10 |