Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Take a Step Back: Evoking Reasoning via Abstraction in Large Language Models

About

We present Step-Back Prompting, a simple prompting technique that enables LLMs to do abstractions to derive high-level concepts and first principles from instances containing specific details. Using the concepts and principles to guide reasoning, LLMs significantly improve their abilities in following a correct reasoning path towards the solution. We conduct experiments of Step-Back Prompting with PaLM-2L, GPT-4 and Llama2-70B models, and observe substantial performance gains on various challenging reasoning-intensive tasks including STEM, Knowledge QA, and Multi-Hop Reasoning. For instance, Step-Back Prompting improves PaLM-2L performance on MMLU (Physics and Chemistry) by 7% and 11% respectively, TimeQA by 27%, and MuSiQue by 7%.

Huaixiu Steven Zheng, Swaroop Mishra, Xinyun Chen, Heng-Tze Cheng, Ed H. Chi, Quoc V Le, Denny Zhou• 2023

Related benchmarks

TaskDatasetResultRank
Mathematical Problem SolvingGaokao MathQA
Accuracy64.5
60
Question AnsweringGPQA (test)
Accuracy42.4
55
Knowledge IntensiveGaokao History
Accuracy76.3
30
Financial Question AnsweringFinanceIQ
Accuracy (%)66.85
27
Sentencing PredictionCAIL Law Domain
Accuracy72.5
24
Logical reasoningGeoShape BBEH
Accuracy25
20
Mathematical CalculationAQUA-RAT
Accuracy (AQuA-RAT)81.8
20
Logical reasoningGeoShape BBH
Accuracy61
20
Prompt OptimizationLogical Reasoning, Mathematical Calculation, and Knowledge Intensive tasks Average
Average Performance (%)59.9
20
Knowledge IntensiveGaokao Geography
Accuracy70.7
20
Showing 10 of 19 rows

Other info

Follow for update