Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Multi-expert Prompting Improves Reliability, Safety, and Usefulness of Large Language Models

About

We present Multi-expert Prompting, a novel enhancement of ExpertPrompting (Xu et al., 2023), designed to improve the large language model (LLM) generation. Specifically, it guides an LLM to fulfill an input instruction by simulating multiple experts, aggregating their responses, and selecting the best among individual and aggregated responses. This process is performed in a single chain of thoughts through our seven carefully designed subtasks derived from the Nominal Group Technique (Ven and Delbecq, 1974), a well-established decision-making framework. Our evaluations demonstrate that Multi-expert Prompting significantly outperforms ExpertPrompting and comparable baselines in enhancing the truthfulness, factuality, informativeness, and usefulness of responses while reducing toxicity and hurtfulness. It further achieves state-of-the-art truthfulness by outperforming the best baseline by 8.69% with ChatGPT. Multi-expert Prompting is efficient, explainable, and highly adaptable to diverse scenarios, eliminating the need for manual prompt construction.

Do Xuan Long, Duong Ngoc Yen, Anh Tuan Luu, Kenji Kawaguchi, Min-Yen Kan, Nancy F. Chen• 2024

Related benchmarks

TaskDatasetResultRank
TruthfulnessTruthfulQA
Truthfulness Accuracy89.35
86
Toxicity EvaluationBoLD
Toxic Rate0.00e+0
26
Logical CoherenceCIAR
Accuracy82
12
Safety EvaluationHONEST
Score0.7
12
Showing 4 of 4 rows

Other info

Follow for update