Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Positive-Unlabeled Learning with Non-Negative Risk Estimator

About

From only positive (P) and unlabeled (U) data, a binary classifier could be trained with PU learning, in which the state of the art is unbiased PU learning. However, if its model is very flexible, empirical risks on training data will go negative, and we will suffer from serious overfitting. In this paper, we propose a non-negative risk estimator for PU learning: when getting minimized, it is more robust against overfitting, and thus we are able to use very flexible models (such as deep neural networks) given limited P data. Moreover, we analyze the bias, consistency, and mean-squared-error reduction of the proposed risk estimator, and bound the estimation error of the resulting empirical risk minimizer. Experiments demonstrate that our risk estimator fixes the overfitting problem of its unbiased counterparts.

Ryuichi Kiryo, Gang Niu, Marthinus C. du Plessis, Masashi Sugiyama• 2017

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-10 (test)
Accuracy84.7
3381
Image ClassificationSTL-10 (test)
Accuracy80.4
357
Out-of-Distribution DetectionCIFAR-10 (ID) vs SVHN (OOD) (test)
AUROC100
88
Image ClassificationF-MNIST (test)
Accuracy91.4
64
Reward ModelingHelpSteer (test)
MAE0.2688
48
Reward ModelingUltraFeedback (test)
MAE0.2021
21
Reward ModelingPKU-SafeRLHF (test)
MAE0.1279
19
Positive-Unlabeled ClassificationCIFAR-10 (test)
Accuracy88.91
19
PvN classificationBinarized CIFAR
Accuracy77.2
18
Outlier DetectionIoT (test)
AUC91.18
17
Showing 10 of 58 rows

Other info

Follow for update