Achieving Fairness at No Utility Cost via Data Reweighing with Influence
About
With the fast development of algorithmic governance, fairness has become a compulsory property for machine learning models to suppress unintentional discrimination. In this paper, we focus on the pre-processing aspect for achieving fairness, and propose a data reweighing approach that only adjusts the weight for samples in the training phase. Different from most previous reweighing methods which usually assign a uniform weight for each (sub)group, we granularly model the influence of each training sample with regard to fairness-related quantity and predictive utility, and compute individual weights based on influence under the constraints from both fairness and utility. Experimental results reveal that previous methods achieve fairness at a non-negligible cost of utility, while as a significant advantage, our approach can empirically release the tradeoff and obtain cost-free fairness for equal opportunity. We demonstrate the cost-free fairness through vanilla classifiers and standard training processes, compared to baseline methods on multiple real-world tabular datasets. Code available at https://github.com/brandeis-machine-learning/influence-fairness.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Classification | Adult | Accuracy82.6 | 21 | |
| Classification | COMM | Accuracy81.95 | 20 | |
| Classification | German | Delta DP0.0054 | 20 | |
| Fair Classification | Adult | Delta DP-0.0504 | 16 | |
| Fair Classification | COMPAS | DP Disparity0.1188 | 16 | |
| Fair Classification | COMM | Delta DP0.337 | 15 | |
| Classification | COMPAS | Accuracy64.96 | 15 | |
| Classification | German | Acc66 | 15 | |
| Classification | Adult | Delta DP15.96 | 7 | |
| Classification | COMPAS | Delta DP0.103 | 7 |