Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Chameleon: Plug-and-Play Compositional Reasoning with Large Language Models

About

Large language models (LLMs) have achieved remarkable progress in solving various natural language processing tasks due to emergent reasoning abilities. However, LLMs have inherent limitations as they are incapable of accessing up-to-date information (stored on the Web or in task-specific knowledge bases), using external tools, and performing precise mathematical and logical reasoning. In this paper, we present Chameleon, an AI system that mitigates these limitations by augmenting LLMs with plug-and-play modules for compositional reasoning. Chameleon synthesizes programs by composing various tools (e.g., LLMs, off-the-shelf vision models, web search engines, Python functions, and heuristic-based modules) for accomplishing complex reasoning tasks. At the heart of Chameleon is an LLM-based planner that assembles a sequence of tools to execute to generate the final response. We showcase the effectiveness of Chameleon on two multi-modal knowledge-intensive reasoning tasks: ScienceQA and TabMWP. Chameleon, powered by GPT-4, achieves an 86.54% overall accuracy on ScienceQA, improving the best published few-shot result by 11.37%. On TabMWP, GPT-4-powered Chameleon improves the accuracy by 17.0%, lifting the state of the art to 98.78%. Our analysis also shows that the GPT-4-powered planner exhibits more consistent and rational tool selection via inferring potential constraints from instructions, compared to a ChatGPT-powered planner. The project is available at https://chameleon-llm.github.io.

Pan Lu, Baolin Peng, Hao Cheng, Michel Galley, Kai-Wei Chang, Ying Nian Wu, Song-Chun Zhu, Jianfeng Gao• 2023

Related benchmarks

TaskDatasetResultRank
Science Question AnsweringScienceQA (test)
Average Accuracy86.54
245
Visual Question AnsweringA-OKVQA
Acc47.5
202
Massive Multi-discipline Multimodal UnderstandingMMMU
Accuracy51.6
152
Science Question AnsweringScienceQA
IMG Score71.49
64
Multimodal ReasoningM3CoT (test)
Total Acc34.29
47
Tool CallingAPI-Bank L-1--
46
Multimodal Science Question AnsweringScienceQA
Overall Average Score83.99
36
Tool LearningRestBench TMDB
Success Rate75
32
Multimodal Science Question AnsweringScienceQA v1.0 (test)
Accuracy (Natural Language Component)89.83
31
Science Question AnsweringScienceQA v1.0 (test)
Accuracy (G1-4)77.83
26
Showing 10 of 26 rows

Other info

Code

Follow for update