A Simple Interpretable Transformer for Fine-Grained Image Classification and Analysis
About
We present a novel usage of Transformers to make image classification interpretable. Unlike mainstream classifiers that wait until the last fully connected layer to incorporate class information to make predictions, we investigate a proactive approach, asking each class to search for itself in an image. We realize this idea via a Transformer encoder-decoder inspired by DEtection TRansformer (DETR). We learn "class-specific" queries (one for each class) as input to the decoder, enabling each class to localize its patterns in an image via cross-attention. We name our approach INterpretable TRansformer (INTR), which is fairly easy to implement and exhibits several compelling properties. We show that INTR intrinsically encourages each class to attend distinctively; the cross-attention weights thus provide a faithful interpretation of the prediction. Interestingly, via "multi-head" cross-attention, INTR could identify different "attributes" of a class, making it particularly suitable for fine-grained classification and analysis, which we demonstrate on eight datasets. Our code and pre-trained models are publicly accessible at the Imageomics Institute GitHub site: https://github.com/Imageomics/INTR.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Hierarchical classification | FathomNet Private 2025 (test) | Hierarchical Distance (HD)3.83 | 15 | |
| Hierarchical classification | FathomNet Public 2025 (test) | Hierarchical Distance (HD)3.53 | 15 | |
| Hierarchical classification | FathomNet Weighted Overall 2025 (Weighted Public Private) | Weighted Hierarchical Distance (WgtAvg)3.68 | 15 | |
| Classification | FishCLEF 2015 | Accuracy67.4 | 10 | |
| Classification | FAIR1M domain generalization evaluation v2 | Accuracy (ACC)63.3 | 10 |