Think-Augmented Function Calling: Improving LLM Parameter Accuracy Through Embedded Reasoning
About
Large language models (LLMs) have demonstrated remarkable capabilities in function calling for autonomous agents, yet current mechanisms lack explicit reasoning transparency during parameter generation, particularly for complex functions with interdependent parameters. While existing approaches like chain-of-thought prompting operate at the agent level, they fail to provide fine-grained reasoning guidance for individual function parameters. To address these limitations, we propose Think-Augmented Function Calling (TAFC), a novel framework that enhances function calling accuracy through explicit reasoning at both function and parameter levels. Our method introduces a universal "think" parameter augmentation that enables models to articulate their decision-making process, with dynamic optimization for parameter descriptions to improve reasoning quality. For complex parameters, TAFC automatically triggers granular reasoning based on complexity scoring, ensuring appropriate justification for critical decisions. Additionally, we propose reasoning-guided optimization to align generated reasoning with human expectations. TAFC requires no architectural modifications to existing LLMs while maintaining full API compatibility. Evaluation on ToolBench across proprietary and open-source models demonstrates significant improvements in parameter generation accuracy and reasoning coherence for multi-parameter functions, while providing enhanced interpretability for debugging AI agent behaviors.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Function Calling | ToolBench I1-Inst | Pass Rate57.1 | 14 | |
| Function Calling | ToolBench I2-Inst | Pass Rate71.4 | 14 | |
| Function Calling | ToolBench I3-Inst | Pass Rate52.4 | 14 | |
| Function Calling | ToolBench Average | Pass Rate60.3 | 14 | |
| Tool-use Parameter Quality Assessment | ToolBench | TAFC Win Rate76.1 | 8 |