Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Faithful Reasoning Using Large Language Models

About

Although contemporary large language models (LMs) demonstrate impressive question-answering capabilities, their answers are typically the product of a single call to the model. This entails an unwelcome degree of opacity and compromises performance, especially on problems that are inherently multi-step. To address these limitations, we show how LMs can be made to perform faithful multi-step reasoning via a process whose causal structure mirrors the underlying logical structure of the problem. Our approach works by chaining together reasoning steps, where each step results from calls to two fine-tuned LMs, one for selection and one for inference, to produce a valid reasoning trace. Our method carries out a beam search through the space of reasoning traces to improve reasoning quality. We demonstrate the effectiveness of our model on multi-step logical deduction and scientific question-answering, showing that it outperforms baselines on final answer accuracy, and generates humanly interpretable reasoning traces whose validity can be checked by the user.

Antonia Creswell, Murray Shanahan• 2022

Related benchmarks

TaskDatasetResultRank
Question AnsweringEntailmentBankQA Task 1 (test)
Accuracy83.2
7
Question AnsweringEntailmentBankQA Task 2 (test)
Accuracy72.9
7
Showing 2 of 2 rows

Other info

Follow for update