Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Answering Questions by Meta-Reasoning over Multiple Chains of Thought

About

Modern systems for multi-hop question answering (QA) typically break questions into a sequence of reasoning steps, termed chain-of-thought (CoT), before arriving at a final answer. Often, multiple chains are sampled and aggregated through a voting mechanism over the final answers, but the intermediate steps themselves are discarded. While such approaches improve performance, they do not consider the relations between intermediate steps across chains and do not provide a unified explanation for the predicted answer. We introduce Multi-Chain Reasoning (MCR), an approach which prompts large language models to meta-reason over multiple chains of thought, rather than aggregating their answers. MCR examines different reasoning chains, mixes information between them and selects the most relevant facts in generating an explanation and predicting the answer. MCR outperforms strong baselines on 7 multi-hop QA datasets. Moreover, our analysis reveals that MCR explanations exhibit high quality, enabling humans to verify its answers.

Ori Yoran, Tomer Wolfson, Ben Bogin, Uri Katz, Daniel Deutch, Jonathan Berant• 2023

Related benchmarks

TaskDatasetResultRank
Multi-hop Question AnsweringHotpotQA--
221
Multi-hop Question Answering2WikiMQA--
154
Question AnsweringStrategyQA
Accuracy73.6
114
Multi-hop Question AnsweringHotpotQA (dev)--
43
Question AnsweringStrategyQA (test)
Task Accuracy75.3
28
Multi-hop Question Answering2WikiMultiHopQA (dev)
Exact Match Accuracy68.6
11
Multi-hop Open-domain Question AnsweringFermi
Accuracy38.9
6
Multi-hop Open-domain Question AnsweringQuaRTz
Accuracy81.6
6
Multi-hop Open-domain Question AnsweringBamboogle
Accuracy66.5
6
Multi-hop Open-domain Question AnsweringFEVEROUS
Accuracy0.694
6
Showing 10 of 12 rows

Other info

Code

Follow for update