Try, Check and Retry: A Divide-and-Conquer Framework for Boosting Long-context Tool-Calling Performance of LLMs
About
Tool-calling empowers Large Language Models (LLMs) to interact with external environments. However, current methods often struggle to handle massive and noisy candidate tools in long-context tool-calling tasks, limiting their real-world application. To this end, we propose Tool-DC, a Divide-and-Conquer framework for boosting tool-calling performance of LLMs. The core of Tool-DC is to reduce the reasoning difficulty and make full use of self-reflection ability of LLMs via a "Try-Check-Retry" paradigm. Specifically, Tool-DC involves two variants: 1) the training-free Tool-DC (TF), which is plug-and-play and flexible; 2) the training-based Tool-DC (TB), which is more inference-efficient. Extensive experiments show that both Tool-DC methods outperform their counterparts by a clear margin. Tool-DC (TF) brings up to +25.10% average gains against the baseline on BFCL and ACEBench benchmarks, while Tool-DC (TB) enables Qwen2.5-7B to achieve comparable or even better performance than proprietary LLMs, e.g., OpenAI o3 and Claude-Haiku-4.5.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Tool Calling | BFCL Extended Setting | Non-Live Score84.4 | 18 | |
| Tool Calling | ACEBench Standard Setting | Overall Score64 | 18 | |
| Tool Calling | ACEBench Extended Setting | Overall Score58.83 | 18 | |
| Tool Calling | BFCL Standard Setting | Non-Live Accuracy84.73 | 18 | |
| Tool Calling | BFCL | Non-Live Success Rate86.67 | 17 | |
| Function Calling | Berkeley Function Calling Leaderboard (BFCL) Extended Setting (Non-Live) | Simple Success Rate73.17 | 6 | |
| Function Calling | Berkeley Function Calling Leaderboard (BFCL) Extended Setting (Live) | Simple Success Rate65.89 | 6 | |
| Function Calling | Berkeley Function Calling Leaderboard (BFCL) Extended Setting (Overall) | Non-Live Score82.54 | 6 |