Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Take a Step Back: Evoking Reasoning via Abstraction in Large Language Models

About

We present Step-Back Prompting, a simple prompting technique that enables LLMs to do abstractions to derive high-level concepts and first principles from instances containing specific details. Using the concepts and principles to guide reasoning, LLMs significantly improve their abilities in following a correct reasoning path towards the solution. We conduct experiments of Step-Back Prompting with PaLM-2L, GPT-4 and Llama2-70B models, and observe substantial performance gains on various challenging reasoning-intensive tasks including STEM, Knowledge QA, and Multi-Hop Reasoning. For instance, Step-Back Prompting improves PaLM-2L performance on MMLU (Physics and Chemistry) by 7% and 11% respectively, TimeQA by 27%, and MuSiQue by 7%.

Huaixiu Steven Zheng, Swaroop Mishra, Xinyun Chen, Heng-Tze Cheng, Ed H. Chi, Quoc V Le, Denny Zhou• 2023

Related benchmarks

TaskDatasetResultRank
Question AnsweringGPQA (test)
Accuracy42.4
55
Financial Question AnsweringFinanceIQ
Accuracy (%)66.85
27
Sentencing PredictionCAIL Law Domain
Accuracy72.5
24
STEM Task EvaluationMMLU Math
Accuracy45.65
18
STEM Task EvaluationMMLU Physics
Accuracy34.31
18
STEM Task EvaluationMMLU Biology
Accuracy0.5347
18
Mathematical ReasoningAGIEval-MATH (test)
Accuracy47.5
11
Coreference ResolutionWSC (test)
Accuracy78.7
11
Navigation ReasoningBBH-Navigate (test)
Accuracy93.5
11
Fact CheckingLIAR (test)
Accuracy62.8
11
Showing 10 of 10 rows

Other info

Follow for update