Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Multi-Persona Thinking for Bias Mitigation in Large Language Models

About

Large Language Models (LLMs) exhibit significant social biases that can perpetuate harmful stereotypes and unfair outcomes. In this paper, we propose Multi-Persona Thinking (MPT), a novel inference-time framework that leverages dialectical reasoning from multiple perspectives to reduce bias. MPT guides models to adopt contrasting social identities (e.g., male and female) along with a neutral viewpoint, and then engages these personas iteratively to expose and correct biases. Through a dialectical reasoning process, the framework transforms the potential weakness of persona assignment into a strength for bias mitigation. We evaluate MPT on two widely used bias benchmarks across both open-source and closed-source models of varying scales. Our results demonstrate substantial improvements over existing prompting-based strategies: MPT achieves the lowest bias while maintaining core reasoning ability.

Yuxing Chen, Guoqing Luo, Zijun Wu, Lili Mou• 2026

Related benchmarks

TaskDatasetResultRank
Question AnsweringBBQ--
36
Question AnsweringBBQ (test)
Accuracy (amb)98.46
20
Showing 2 of 2 rows

Other info

Follow for update