Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Debug like a Human: A Large Language Model Debugger via Verifying Runtime Execution Step-by-step

About

Large language models (LLMs) are leading significant progress in code generation. Beyond one-pass code generation, recent works further integrate unit tests and program verifiers into LLMs to iteratively refine the generated programs. However, these works consider the generated programs as an indivisible entity, which falls short for LLMs in debugging the programs, especially when the programs contain complex logic flows and data operations. In contrast, when human developers debug programs, they typically set breakpoints and selectively examine runtime execution information. The execution flow and the intermediate variables play a crucial role in the debugging process, yet they are underutilized in the existing literature on code generation. In this study, we introduce Large Language Model Debugger (LDB), a novel debugging framework that enables LLMs to refine their generated programs with the runtime execution information. Specifically, LDB segments the programs into basic blocks and tracks the values of intermediate variables after each block throughout the runtime execution. This allows LLMs to concentrate on simpler code units within the overall execution flow, verify their correctness against the task description block by block, and efficiently pinpoint any potential errors. Experiments demonstrate that LDB consistently enhances the baseline performance by up to 9.8% across the HumanEval, MBPP, and TransCoder benchmarks, archiving new state-of-the-art performance in code debugging for various LLM selections.

Li Zhong, Zilong Wang, Jingbo Shang• 2024

Related benchmarks

TaskDatasetResultRank
Code GenerationHumanEval
Pass@198.2
850
Code GenerationHumanEval (test)
Pass@182.9
444
Code GenerationMBPP (test)
Pass@176
276
Code GenerationHumanEval 1.0 (test)
Pass@10.829
145
Code GenerationHumanEval+ (test)--
81
Code GenerationHumanEval-ET
Pass@181.7
75
Code GenerationMBPP-ET
Pass@165.4
75
Code DebuggingHumanEval
Accuracy84.1
42
Code GenerationCodeContests
Pass@129.3
38
Code GenerationAPPS Intermediate
Pass Rate46.78
32
Showing 10 of 22 rows

Other info

Code

Follow for update