Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Certifying Robustness to Programmable Data Bias in Decision Trees

About

Datasets can be biased due to societal inequities, human biases, under-representation of minorities, etc. Our goal is to certify that models produced by a learning algorithm are pointwise-robust to potential dataset biases. This is a challenging problem: it entails learning models for a large, or even infinite, number of datasets, ensuring that they all produce the same prediction. We focus on decision-tree learning due to the interpretable nature of the models. Our approach allows programmatically specifying bias models across a variety of dimensions (e.g., missing data for minorities), composing types of bias, and targeting bias towards a specific group. To certify robustness, we use a novel symbolic technique to evaluate a decision-tree learner on a large, or infinite, number of datasets, certifying that each and every dataset produces the same prediction for a specific test point. We evaluate our approach on datasets that are commonly used in the fairness literature, and demonstrate our approach's viability on a range of bias models.

Anna P. Meyer, Aws Albarghouthi, Loris D'Antoni• 2021

Related benchmarks

TaskDatasetResultRank
Binary ClassificationMNIST 0 and 1 (test)
Certification Rate100
24
Robustness CertificationCOMPAS (test)
Certification Rate (0.05% bias)89
6
Robustness CertificationAdult Income (AI) (test)
Certification Rate (0.05% Bias)98.8
6
Robustness CertificationDrug Consumption (test)
Certification Rate (0.05% bias)94.5
4
Showing 4 of 4 rows

Other info

Code

Follow for update