Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Input Similarity from the Neural Network Perspective

About

We first exhibit a multimodal image registration task, for which a neural network trained on a dataset with noisy labels reaches almost perfect accuracy, far beyond noise variance. This surprising auto-denoising phenomenon can be explained as a noise averaging effect over the labels of similar input examples. This effect theoretically grows with the number of similar examples; the question is then to define and estimate the similarity of examples. We express a proper definition of similarity, from the neural network perspective, i.e. we quantify how undissociable two inputs $A$ and $B$ are, taking a machine learning viewpoint: how much a parameter variation designed to change the output for $A$ would impact the output for $B$ as well? We study the mathematical properties of this similarity measure, and show how to use it on a trained network to estimate sample density, in low complexity, enabling new types of statistical analysis for neural networks. We analyze data by retrieving samples perceived as similar by the network, and are able to quantify the denoising effect without requiring true labels. We also propose, during training, to enforce that examples known to be similar should also be seen as similar by the network, and notice speed-up training effects for certain datasets.

Guillaume Charpiat, Nicolas Girard, Loris Felardos, Yuliya Tarabalka• 2021

Related benchmarks

TaskDatasetResultRank
Contributor AttributionFashion Product
Diversity19.41
48
Contributor AttributionArtBench Post-Impressionism
Aesthetic Score6.92
36
Contributor AttributionCIFAR-20
Inception Score18.69
32
Contributor AttributionArtBench Post-Impressionism (test)
Aesthetic Score0.34
18
Contributor AttributionCIFAR-20 (test)
Inception Score29.44
16
Training Data AttributionGPT2-small
LDS Score0.0103
10
Training Data AttributionOlmo-7B
Tail-patch (%)14.4
5
Showing 7 of 7 rows

Other info

Follow for update