Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

From LLMs to Actions: Latent Codes as Bridges in Hierarchical Robot Control

About

Hierarchical control for robotics has long been plagued by the need to have a well defined interface layer to communicate between high-level task planners and low-level policies. With the advent of LLMs, language has been emerging as a prospective interface layer. However, this has several limitations. Not all tasks can be decomposed into steps that are easily expressible in natural language (e.g. performing a dance routine). Further, it makes end-to-end finetuning on embodied data challenging due to domain shift and catastrophic forgetting. We introduce our method -- Learnable Latent Codes as Bridges (LCB) -- as an alternate architecture to overcome these limitations. \method~uses a learnable latent code to act as a bridge between LLMs and low-level policies. This enables LLMs to flexibly communicate goals in the task plan without being entirely constrained by language limitations. Additionally, it enables end-to-end finetuning without destroying the embedding space of word tokens learned during pre-training. Through experiments on Language Table and Calvin, two common language based benchmarks for embodied agents, we find that \method~outperforms baselines (including those w/ GPT-4V) that leverage pure language as the interface layer on tasks that require reasoning and multi-step behaviors.

Yide Shentu, Philipp Wu, Aravind Rajeswaran, Pieter Abbeel• 2024

Related benchmarks

TaskDatasetResultRank
Long-horizon task completionCalvin ABC->D
Success Rate (1)73.6
67
Robot ManipulationCALVIN ABC->D 1.0
Success Rate (1 Inst)73.6
18
Robot ManipulationLanguage Table
Average Success Rate80
6
Language-conditioned long-horizon robot manipulationCALVIN GPT-4 enriched Long-horizon (Split D)
Success Rate (1/5)73.6
3
Showing 4 of 4 rows

Other info

Follow for update