Interpreting Neurons in Deep Vision Networks with Language Models
About
In this paper, we propose Describe-and-Dissect (DnD), a novel method to describe the roles of hidden neurons in vision networks. DnD utilizes recent advancements in multimodal deep learning to produce complex natural language descriptions, without the need for labeled training data or a predefined set of concepts to choose from. Additionally, DnD is training-free, meaning we don't train any new models and can easily leverage more capable general purpose models in the future. We have conducted extensive qualitative and quantitative analysis to show that DnD outperforms prior work by providing higher quality neuron descriptions. Specifically, our method on average provides the highest quality labels and is more than 2$\times$ as likely to be selected as the best explanation for a neuron than the best baseline. Finally, we present a use case providing critical insights into land cover prediction models for sustainability applications. Our code and data are available at https://github.com/Trustworthy-ML-Lab/Describe-and-Dissect.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Neuron Interpretation | ImageNet-1k (val) | CLIP Cosine Similarity0.7595 | 18 | |
| Neuron Interpretation | ImageNet CoSy benchmark avgpool layer 1k | AUC0.76 | 12 | |
| Neuron Interpretation | Places365 (test) | CLIP Cosine Similarity0.7458 | 7 | |
| Concept Discovery | ImageNet | AUC10 | 5 | |
| Neuron Interpretation | Places365 CoSy benchmark avgpool layer | AUC74 | 4 |