Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Mitigating Boundary Ambiguity and Inherent Bias for Text Classification in the Era of Large Language Models

About

Text classification is a crucial task encountered frequently in practical scenarios, yet it is still under-explored in the era of large language models (LLMs). This study shows that LLMs are vulnerable to changes in the number and arrangement of options in text classification. Our extensive empirical analyses reveal that the key bottleneck arises from ambiguous decision boundaries and inherent biases towards specific tokens and positions. To mitigate these issues, we make the first attempt and propose a novel two-stage classification framework for LLMs. Our approach is grounded in the empirical observation that pairwise comparisons can effectively alleviate boundary ambiguity and inherent bias. Specifically, we begin with a self-reduction technique to efficiently narrow down numerous options, which contributes to reduced decision space and a faster comparison process. Subsequently, pairwise contrastive comparisons are employed in a chain-of-thought manner to draw out nuances and distinguish confusable options, thus refining the ambiguous decision boundary. Extensive experiments on four datasets (Banking77, HWU64, LIU54, and Clinic150) verify the effectiveness of our framework. Furthermore, benefitting from our framework, various LLMs can achieve consistent improvements. Our code and data are available in \url{https://github.com/Chuge0335/PC-CoT}.

Zhenyi Lu, Jie Tian, Wei Wei, Xiaoye Qu, Yu Cheng, Wenfeng xie, Dangyang Chen• 2024

Related benchmarks

TaskDatasetResultRank
Intent ClassificationBanking77 (test)
Accuracy61.82
151
Intent ClassificationClinc150 (test)
Accuracy84.89
26
Intent ClassificationHWU64 (test)
Accuracy65.09
9
Intent ClassificationLIU54 (test)
Accuracy41.73
9
Showing 4 of 4 rows

Other info

Code

Follow for update