Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

OMP: One-step Meanflow Policy with Directional Alignment

About

Robot manipulation has increasingly adopted data-driven generative policy frameworks, yet the field faces a persistent trade-off: diffusion models suffer from high inference latency, while flow-based methods often require complex architectural constraints. Although in image generation domain, the MeanFlow paradigm offers a path to single-step inference, its direct application to robotics is impeded by critical theoretical pathologies, specifically spectral bias and gradient starvation in low-velocity regimes. To overcome these limitations, we propose the One-step MeanFlow Policy (OMP), a novel framework designed for high-fidelity, real-time manipulation. We introduce a lightweight directional alignment mechanism to explicitly synchronize predicted velocities with true mean velocities. Furthermore, we implement a Differential Derivation Equation (DDE) to approximate the Jacobian-Vector Product (JVP) operator, which decouples forward and backward passes to significantly reduce memory complexity. Extensive experiments on the Adroit and Meta-World benchmarks demonstrate that OMP outperforms state-of-the-art methods in success rate and trajectory accuracy, particularly in high-precision tasks, while retaining the efficiency of single-step generation.

Han Fang, Yize Huang, Yuheng Zhao, Paul Weng, Xiao Li, Yutong Ban• 2025

Related benchmarks

TaskDatasetResultRank
Robot ManipulationAdroit
Pen Task Score64
50
3D pointcloud manipulationMetaWorld
Success Rate (Easy)89.7
17
Robot ManipulationMeta-World
Latency (Easy) (ms)87.5
15
Robot Manipulation (Clean Table)Real-world robot experiments
Success Rate75
4
Robot Manipulation (Place Bottle)Real-world robot experiments
Success Rate80
4
Robot Manipulation (Slip Ring)Real-world robot experiments
Success Rate70
4
Showing 6 of 6 rows

Other info

Follow for update