Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Preserving Fairness and Safety in Quantized LLMs Through Critical Weight Protection

About

Quantization is widely adopted to reduce the computational cost of large language models (LLMs); however, its implications for fairness and safety, particularly in dynamic quantization and multilingual contexts, remain underexplored. In this work, we conduct a systematic study of how static and dynamic quantization methods impact fairness and safety across benchmarks measuring intrinsic and extrinsic bias and safety alignment. For fairness, we evaluate English, French, Dutch, Spanish, and Turkish; for safety, we focus on English, Korean, and Arabic. Our findings reveal that quantization consistently degrades fairness and safety, with dynamic methods demonstrating greater stability than static ones. Moreover, fairness degradation varies across languages, while safety deterioration is especially pronounced in non-English settings. To address these risks, we introduce Critical Weight Protection, a novel technique that identifies and preserves fairness- and safety-critical weights during quantization. This approach effectively mitigates bias and safety deterioration without costly retraining or alignment, maintaining trustworthiness while retaining efficiency.

Muhammad Alif Al Hakim, Alfan Farizki Wicaksono, Fajri Koto• 2026

Related benchmarks

TaskDatasetResultRank
Bias MeasurementStereoSet
Overall SS50.215
25
Safety EvaluationHEX-PHI (test)
ASR7.333
12
Fairness evaluationJigsaw
BiasAUC75.6
9
Safety EvaluationSafetyBench (test)
Accuracy81.321
9
Safety EvaluationMultiJail-KO (test)
Safety Rate90.053
9
Safety EvaluationMultiJail-AR (test)
Safety Rate (%)89.101
9
Fairness evaluationCrowS-Pair En
Stereotype Score64.58
9
Safety EvaluationDo-Not-Answer (test)
ASR3.301
9
Fairness evaluationCrowS-Pair Fr
SS52.057
9
Safety EvaluationMultiJail EN (test)
Safety Rate92.487
9
Showing 10 of 10 rows

Other info

Follow for update