Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

CLIP-Dissect: Automatic Description of Neuron Representations in Deep Vision Networks

About

In this paper, we propose CLIP-Dissect, a new technique to automatically describe the function of individual hidden neurons inside vision networks. CLIP-Dissect leverages recent advances in multimodal vision/language models to label internal neurons with open-ended concepts without the need for any labeled data or human examples. We show that CLIP-Dissect provides more accurate descriptions than existing methods for last layer neurons where the ground-truth is available as well as qualitatively good descriptions for hidden layer neurons. In addition, our method is very flexible: it is model agnostic, can easily handle new concepts and can be extended to take advantage of better multimodal models in the future. Finally CLIP-Dissect is computationally efficient and can label all neurons from five layers of ResNet-50 in just 4 minutes, which is more than 10 times faster than existing methods. Our code is available at https://github.com/Trustworthy-ML-Lab/CLIP-dissect. Finally, crowdsourced user study results are available at Appendix B to further support the effectiveness of our method.

Tuomas Oikarinen, Tsui-Wei Weng• 2022

Related benchmarks

TaskDatasetResultRank
Neuron descriptionImageNet
AUC95
15
Neuron IdentificationBroden
Accuracy70.5
6
Neuron descriptionPlaces365
AUC0.92
6
Neuron IdentificationImageNet Broden (val)
Accuracy95.4
6
Neuron IdentificationImageNet (val)
Accuracy95
6
Neuron IdentificationCIFAR-100 (train)
Accuracy46.2
6
Concept Activation AnalysisADE20K (test)
CAV Mean91.6
3
Concept LabelingMedical Diagnosis (test)
Count (90-100%)17
3
Neuron Concept AnnotationNIH14 (test)
Cosine Similarity (CLIP)0.8205
3
Hidden Neuron Activation AnalysisHidden Neuron Activations
Count (90-100% Activation Bin)4
3
Showing 10 of 10 rows

Other info

Follow for update