Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Deep Generative Views to Mitigate Gender Classification Bias Across Gender-Race Groups

About

Published studies have suggested the bias of automated face-based gender classification algorithms across gender-race groups. Specifically, unequal accuracy rates were obtained for women and dark-skinned people. To mitigate the bias of gender classifiers, the vision community has developed several strategies. However, the efficacy of these mitigation strategies is demonstrated for a limited number of races mostly, Caucasian and African-American. Further, these strategies often offer a trade-off between bias and classification accuracy. To further advance the state-of-the-art, we leverage the power of generative views, structured learning, and evidential learning towards mitigating gender classification bias. We demonstrate the superiority of our bias mitigation strategy in improving classification accuracy and reducing bias across gender-racial groups through extensive experimental validation, resulting in state-of-the-art performance in intra- and cross dataset evaluations.

Sreeraj Ramachandran, Ajita Rattani• 2022

Related benchmarks

TaskDatasetResultRank
Gender ClassificationFairFace v1 (test)
Accuracy (Black)95.37
20
Gender ClassificationUTKFace Cross-dataset
DoB1.6
9
Gender ClassificationDiveFace (Cross-dataset)
DoB0.83
9
Gender ClassificationMorph (Cross-dataset)
DoB1.1
9
Gender ClassificationVISOB
Avg. Acc0.8917
2
Gender ClassificationUFPR
Accuracy (Male)97.13
2
Showing 6 of 6 rows

Other info

Follow for update