CLIP-Dissect: Automatic Description of Neuron Representations in Deep Vision Networks
About
In this paper, we propose CLIP-Dissect, a new technique to automatically describe the function of individual hidden neurons inside vision networks. CLIP-Dissect leverages recent advances in multimodal vision/language models to label internal neurons with open-ended concepts without the need for any labeled data or human examples. We show that CLIP-Dissect provides more accurate descriptions than existing methods for last layer neurons where the ground-truth is available as well as qualitatively good descriptions for hidden layer neurons. In addition, our method is very flexible: it is model agnostic, can easily handle new concepts and can be extended to take advantage of better multimodal models in the future. Finally CLIP-Dissect is computationally efficient and can label all neurons from five layers of ResNet-50 in just 4 minutes, which is more than 10 times faster than existing methods. Our code is available at https://github.com/Trustworthy-ML-Lab/CLIP-dissect. Finally, crowdsourced user study results are available at Appendix B to further support the effectiveness of our method.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Neuron Interpretation | ImageNet-1k (val) | CLIP Cosine Similarity0.7942 | 18 | |
| Neuron description | ImageNet | AUC95 | 15 | |
| Visual Grounding | ImageNet-1k (val) | Alignment Score0.85 | 14 | |
| Neuron Interpretation | ImageNet CoSy benchmark avgpool layer 1k | AUC0.91 | 12 | |
| Latent Training Semantics Recovery | ImageNet 1k (train) | Semantic Similarity Score73 | 10 | |
| Neuron Interpretation | Places365 (test) | CLIP Cosine Similarity0.7883 | 7 | |
| Neuron Identification | Broden | Accuracy70.5 | 6 | |
| Neuron description | Places365 | AUC0.92 | 6 | |
| Neuron Identification | ImageNet Broden (val) | Accuracy95.4 | 6 | |
| Neuron Identification | ImageNet (val) | Accuracy95 | 6 |