Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Identifying and Correcting Label Bias in Machine Learning

About

Datasets often contain biases which unfairly disadvantage certain groups, and classifiers trained on such datasets can inherit these biases. In this paper, we provide a mathematical formulation of how this bias can arise. We do so by assuming the existence of underlying, unknown, and unbiased labels which are overwritten by an agent who intends to provide accurate labels but may have biases against certain groups. Despite the fact that we only observe the biased labels, we are able to show that the bias may nevertheless be corrected by re-weighting the data points without changing the labels. We show, with theoretical guarantees, that training on the re-weighted dataset corresponds to training on the unobserved but unbiased labels, thus leading to an unbiased machine learning classifier. Our procedure is fast and robust and can be used with virtually any learning algorithm. We evaluate on a number of standard machine learning fairness datasets and a variety of fairness notions, finding that our method outperforms standard approaches in achieving fair classification.

Heinrich Jiang, Ofir Nachum• 2019

Related benchmarks

TaskDatasetResultRank
ClassificationBank
Accuracy70.58
25
ClassificationAdult
Accuracy82.43
21
ClassificationGerman
Delta DP-0.1023
20
ClassificationCOMM
Accuracy79.7
20
ClassificationMEPS
AUC83.06
19
ClassificationLSAC
AUC0.8665
19
Fair ClassificationAdult
Delta DP-0.1598
16
Fair ClassificationCOMPAS
DP Disparity-0.1743
16
Fair ClassificationCOMM
Delta DP0.1731
15
ClassificationCOMPAS
Accuracy65.21
15
Showing 10 of 12 rows

Other info

Follow for update