A Near-Optimal Algorithm for Debiasing Trained Machine Learning Models
About
We present a scalable post-processing algorithm for debiasing trained models, including deep neural networks (DNNs), which we prove to be near-optimal by bounding its excess Bayes risk. We empirically validate its advantages on standard benchmark datasets across both classical algorithms as well as modern DNN architectures and demonstrate that it outperforms previous post-processing methods while performing on par with in-processing. In addition, we show that the proposed algorithm is particularly effective for models trained at scale where post-processing is a natural and practical choice.
Ibrahim Alabdulmohsin, Mario Lucic• 2021
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Classification | Adult (test) | Bias0.01 | 24 | |
| Classification | DCCC (test) | Bias0.01 | 24 | |
| Classification | Adult Income 1996 (train and test) | Demographic Parity2.2 | 24 | |
| Classification | Adult (test) | Min Test Accuracy84.4 | 24 | |
| Attribute Prediction | CelebA (test) | Bias0.002 | 20 | |
| Binary Classification | DCCC (test) | Accuracy (Test)81.8 | 16 | |
| Binary Classification | CelebA (test) | Accuracy72.8 | 12 | |
| Fairness-aware Classification | Adult | Training Time (min)1 | 7 | |
| Fairness-aware Classification | COMPAS | Training Time (min)1 | 7 | |
| Fairness-aware Classification | CelebA | Training Time (min)10 | 7 |
Showing 10 of 14 rows