Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Towards Achieving Concept Completeness for Textual Concept Bottleneck Models

About

Textual Concept Bottleneck Models (TCBMs) are interpretable-by-design models for text classification that predict a set of salient concepts before making the final prediction. This paper proposes Complete Textual Concept Bottleneck Model (CT-CBM), a novel TCBM generator building concept labels in a fully unsupervised manner using a small language model, eliminating both the need for predefined human labeled concepts and LLM annotations. CT-CBM iteratively targets and adds important and identifiable concepts in the bottleneck layer to create a complete concept basis. CT-CBM achieves striking results against competitors in terms of concept basis completeness and concept detection accuracy, offering a promising solution to reliably enhance interpretability of NLP classifiers.

Milan Bhan, Yann Choho, Pierre Moreau, Jean-Noel Vittaut, Nicolas Chesneau, Marie-Jeanne Lesot• 2025

Related benchmarks

TaskDatasetResultRank
Image ClassificationCUB-200 (test)
Accuracy77.8
98
Text ClassificationDBPedia (test)--
40
News ClassificationAG News (test)
Accuracy91.2
34
ClassificationN24 News (test)
Accuracy98.5
10
Showing 4 of 4 rows

Other info

Follow for update