Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Understanding intermediate layers using linear classifier probes

About

Neural network models have a reputation for being black boxes. We propose to monitor the features at every layer of a model and measure how suitable they are for classification. We use linear classifiers, which we refer to as "probes", trained entirely independently of the model itself. This helps us better understand the roles and dynamics of the intermediate layers. We demonstrate how this can be used to develop a better intuition about models and to diagnose potential problems. We apply this technique to the popular models Inception v3 and Resnet-50. Among other things, we observe experimentally that the linear separability of features increase monotonically along the depth of the model.

Guillaume Alain, Yoshua Bengio• 2016

Related benchmarks

TaskDatasetResultRank
Image ClassificationStanford Cars
Accuracy93.8
635
Image ClassificationEuroSAT
Accuracy97
569
Image ClassificationDTD
Accuracy83.5
485
Image ClassificationSUN397
Accuracy77.2
441
Image ClassificationGTSRB
Accuracy86.8
291
Image ClassificationFashionMNIST (test)
Accuracy85.8
260
Image ClassificationPets--
245
Hallucination DetectionTriviaQA (test)
AUC-ROC83.36
183
Natural Language InferenceSNLI
Accuracy71.2
180
Image ClassificationCIFAR100
Average Accuracy88.5
121
Showing 10 of 90 rows
...

Other info

Follow for update