Mitigating Extrinsic Gender Bias for Bangla Classification Tasks
About
In this study, we investigate extrinsic gender bias in Bangla pretrained language models, a largely underexplored area in low-resource languages. To assess this bias, we construct four manually annotated, task-specific benchmark datasets for sentiment analysis, toxicity detection, hate speech detection, and sarcasm detection. Each dataset is augmented using nuanced gender perturbations, where we systematically swap gendered names and terms while preserving semantic content, enabling minimal-pair evaluation of gender-driven prediction shifts. We then propose RandSymKL, a randomized debiasing strategy integrated with symmetric KL divergence and cross-entropy loss to mitigate the bias across task-specific pretrained models. RandSymKL is a refined training approach to integrate these elements in a unified way for extrinsic gender bias mitigation focused on classification tasks. Our approach was evaluated against existing bias mitigation methods, with results showing that our technique not only effectively reduces bias but also maintains competitive accuracy compared to other baseline approaches. To promote further research, we have made both our implementation and datasets publicly available: https://github.com/sajib-kumar/Mitigating-Bangla-Extrinsic-Gender-Bias
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Hate Speech Detection | Bangla | Accuracy88.66 | 15 | |
| Sentiment Analysis | Bangla | Accuracy95.86 | 15 | |
| Toxicity Detection | Bangla | Accuracy90.62 | 15 | |
| Sarcasm Detection | Bangla | Accuracy88.07 | 15 | |
| Sarcasm Detection | Bangla Sarcasm | EOD0.002 | 8 | |
| Sentiment Analysis | Bangla Sentiment Dataset | Accuracy Gap (AG)0.4 | 8 | |
| Text Classification | Bangla Fairness Suite Aggregate | Average EOD0.002 | 8 | |
| Hate Speech Detection | Bangla HateSpeech | EOD Error0.2 | 8 | |
| Toxicity Detection | Bangla Toxicity Dataset | Accuracy Gap (AG)0.48 | 8 | |
| Hate Speech Detection | Bangla HateSpeech Dataset | Accuracy Gap (AG)0.27 | 8 |