Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Graph of Thoughts: Solving Elaborate Problems with Large Language Models

About

We introduce Graph of Thoughts (GoT): a framework that advances prompting capabilities in large language models (LLMs) beyond those offered by paradigms such as Chain-of-Thought or Tree of Thoughts (ToT). The key idea and primary advantage of GoT is the ability to model the information generated by an LLM as an arbitrary graph, where units of information ("LLM thoughts") are vertices, and edges correspond to dependencies between these vertices. This approach enables combining arbitrary LLM thoughts into synergistic outcomes, distilling the essence of whole networks of thoughts, or enhancing thoughts using feedback loops. We illustrate that GoT offers advantages over state of the art on different tasks, for example increasing the quality of sorting by 62% over ToT, while simultaneously reducing costs by >31%. We ensure that GoT is extensible with new thought transformations and thus can be used to spearhead new prompting schemes. This work brings the LLM reasoning closer to human thinking or brain mechanisms such as recurrence, both of which form complex networks.

Maciej Besta, Nils Blach, Ales Kubicek, Robert Gerstenberger, Michal Podstawski, Lukas Gianinazzi, Joanna Gajda, Tomasz Lehmann, Hubert Niewiadomski, Piotr Nyczyk, Torsten Hoefler• 2023

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K
Accuracy95.4
1362
Mathematical ReasoningMATH
Accuracy88.4
882
ReasoningBBH
Accuracy92.1
672
Multi-hop Question AnsweringHotpotQA--
294
Logical reasoningBBH
Accuracy86.2
201
Mathematical ReasoningGSM8K
Math Score90.72
197
Code GenerationMBPP
Pass@173.2
193
Arithmetic ReasoningGSM8K
Accuracy94.5
173
Code GenerationHumanEval
Pass@185.8
171
Mathematical ReasoningGame of 24
Accuracy90
103
Showing 10 of 98 rows
...

Other info

Follow for update