Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

TSLM: Tree-Structured Language Modeling for Divergent Thinking

About

Language models generate reasoning sequentially, preventing them from decoupling irrelevant exploration paths during search. We introduce Tree-Structured Language Modeling (TSLM), which uses special tokens to encode branching structure, enabling models to generate and selectively expand multiple search paths within a single generation process. By training on complete search trees including both successful and failed attempts, TSLM learns to internalize systematic exploration without redundant recomputation of shared prefixes. TSLM achieves robust performance and superior inference efficiency by avoiding the multiple independent forward passes required by external search methods. These results suggest a new paradigm of inference-time scaling for robust reasoning, demonstrating that supervised learning on complete tree-structured traces provides an efficient alternative for developing systematic exploration capabilities in language models.

Doyoung Kim, Jaehyeok Doo, Minjoon Seo• 2026

Related benchmarks

TaskDatasetResultRank
Mathematical Word Problem SolvingGSM8K--
18
Mathematical ReasoningGame of 24
pass@1100
4
Navigation ReasoningGridworld 10x10 (in-domain)
Pass@1100
4
Navigation ReasoningGridworld o.o.d 20x20 (out-of-domain)
Pass@191.5
4
Showing 4 of 4 rows

Other info

Follow for update