Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Certifying Robustness to Programmable Data Bias in Decision Trees

About

Datasets can be biased due to societal inequities, human biases, under-representation of minorities, etc. Our goal is to certify that models produced by a learning algorithm are pointwise-robust to potential dataset biases. This is a challenging problem: it entails learning models for a large, or even infinite, number of datasets, ensuring that they all produce the same prediction. We focus on decision-tree learning due to the interpretable nature of the models. Our approach allows programmatically specifying bias models across a variety of dimensions (e.g., missing data for minorities), composing types of bias, and targeting bias towards a specific group. To certify robustness, we use a novel symbolic technique to evaluate a decision-tree learner on a large, or infinite, number of datasets, certifying that each and every dataset produces the same prediction for a specific test point. We evaluate our approach on datasets that are commonly used in the fairness literature, and demonstrate our approach's viability on a range of bias models.

Anna P. Meyer, Aws Albarghouthi, Loris D'Antoni• 2021

Related benchmarks

TaskDatasetResultRank
Binary ClassificationMNIST 0 and 1 (test)
Certification Rate100
24
Robustness CertificationCOMPAS (test)
Certification Rate (0.05% bias)89
6
Robustness CertificationAdult Income (AI) (test)
Certification Rate (0.05% Bias)98.8
6
Robustness CertificationDrug Consumption (test)
Certification Rate (0.05% bias)94.5
4
Showing 4 of 4 rows

Other info

Code

Follow for update