Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

TaskBench: Benchmarking Large Language Models for Task Automation

About

In recent years, the remarkable progress of large language models (LLMs) has sparked interest in task automation, which involves decomposing complex tasks described by user instructions into sub-tasks and invoking external tools to execute them, playing a central role in autonomous agents. However, there is a lack of systematic and standardized benchmarks to promote the development of LLMs in task automation. To address this, we introduce TaskBench, a comprehensive framework to evaluate the capability of LLMs in task automation. Specifically, task automation can be divided into three critical stages: task decomposition, tool selection, and parameter prediction. To tackle the complexities inherent in these stages, we introduce the concept of Tool Graph to represent decomposed tasks and adopt a back-instruct method to generate high-quality user instructions. We propose TaskEval, a multi-faceted evaluation methodology that assesses LLM performance across these three stages. Our approach combines automated construction with rigorous human verification, ensuring high consistency with human evaluation. Experimental results demonstrate that TaskBench effectively reflects the capabilities of various LLMs in task automation. It provides insights into model performance across different task complexities and domains, pushing the boundaries of what current models can achieve. TaskBench offers a scalable, adaptable, and reliable benchmark for advancing LLM-based autonomous agents.

Yongliang Shen, Kaitao Song, Xu Tan, Wenqi Zhang, Kan Ren, Siyu Yuan, Weiming Lu, Dongsheng Li, Yueting Zhuang• 2023

Related benchmarks

TaskDatasetResultRank
Fact RetrievalAcademicEval Abstract-single
F124.7
13
Fact RetrievalGov Report
F1 Score23.8
13
Fact RetrievalWCEP
F1 Score21.4
13
Fact RetrievalAcademicEval Abstract-multi
F1 Score20.4
13
Fact RetrievalAcademicEval Related-multi
F1 Score18.3
13
ReasoningLooGLE
Reasoning Accuracy17
10
Question AnsweringLooGLE
QA Accuracy3
10
Showing 7 of 7 rows

Other info

Follow for update