Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SIGUA: Forgetting May Make Learning with Noisy Labels More Robust

About

Given data with noisy labels, over-parameterized deep networks can gradually memorize the data, and fit everything in the end. Although equipped with corrections for noisy labels, many learning methods in this area still suffer overfitting due to undesired memorization. In this paper, to relieve this issue, we propose stochastic integrated gradient underweighted ascent (SIGUA): in a mini-batch, we adopt gradient descent on good data as usual, and learning-rate-reduced gradient ascent on bad data; the proposal is a versatile approach where data goodness or badness is w.r.t. desired or undesired memorization given a base learning method. Technically, SIGUA pulls optimization back for generalization when their goals conflict with each other; philosophically, SIGUA shows forgetting undesired memorization can reinforce desired memorization. Experiments demonstrate that SIGUA successfully robustifies two typical base learning methods, so that their performance is often significantly improved.

Bo Han, Gang Niu, Xingrui Yu, Quanming Yao, Miao Xu, Ivor Tsang, Masashi Sugiyama• 2018

Related benchmarks

TaskDatasetResultRank
Image ClassificationFashion MNIST (test)
Accuracy87.64
568
Image ClassificationClothing1M (test)
Accuracy62.89
546
Image ClassificationCIFAR-100 synthetic noise (test)--
61
Activity RecognitionHHAR (test)
Mean F1 Score0.6894
46
Time-series classificationfNIRS (test)
F1 Score0.6737
36
Sleep stage scoringSleep (test)
F1 Score54.28
36
Age EstimationAFAD B
MRAE (%)5.96
33
Image ClassificationMNIST (test)
Accuracy (Symmetric Noise η=0.2)92.31
22
RegressionMSD-B (test)
MRAE (Symmetric, 20%)1.29
17
Age EstimationIMDB-WIKI-B
MRAE1.96
17
Showing 10 of 18 rows

Other info

Follow for update