Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

On Fairness and Calibration

About

The machine learning community has become increasingly concerned with the potential for bias and discrimination in predictive models. This has motivated a growing line of work on what it means for a classification procedure to be "fair." In this paper, we investigate the tension between minimizing error disparity across different population groups while maintaining calibrated probability estimates. We show that calibration is compatible only with a single error constraint (i.e. equal false-negatives rates across groups), and show that any algorithm that satisfies this relaxation is no better than randomizing a percentage of predictions for an existing classifier. These unsettling findings, which extend and generalize existing results, are empirically confirmed on several datasets.

Geoff Pleiss, Manish Raghavan, Felix Wu, Jon Kleinberg, Kilian Q. Weinberger• 2017

Related benchmarks

TaskDatasetResultRank
Binary ClassificationHeart
Accuracy91.65
12
Binary ClassificationCOMPAS
Accuracy64.77
12
Binary ClassificationGerman
Accuracy68
12
Binary ClassificationAdult
Accuracy83.27
12
Binary ClassificationHeart Disease Cleveland standard processed (train)
AUC0.8314
6
Fairness-aware ClassificationPleiss setup (test)
AUC0.8149
3
Showing 6 of 6 rows

Other info

Follow for update