FRMD: fast robot motion diffusion via trajectory-level consistency distillation

FRMD:基于轨迹级一致性蒸馏的快速机器人运动扩散

阅读:1

Abstract

Foundation models for embodied artificial intelligence (Embodied AI) increasingly adopt diffusion modules as the action generation core of vision-language-action (VLA) policies, but the diffusion module's iterative denoising imposes prohibitive inference latency for real-time deployment. We address this bottleneck in isolation by rethinking the diffusion action generation module itself. We present Fast Robot Motion Diffusion (FRMD) , a fast robot motion diffusion framework that (i) operates in trajectory-parameter space by predicting movement-primitive coefficients in a low-dimensional manifold, and (ii) collapses multi-step sampling into a single inference step via trajectory-level consistency distillation over the probability-flow ordinary differential equation (ODE). Concretely, FRMD replaces stepwise action generation with a one-pass mapping from noise to full trajectories, followed by a fixed-cost basis expansion; this reduces policy latency from hundreds to tens of milliseconds without modifying upstream vision or language encoders. On standard robotic manipulation task benchmarks, FRMD attains 7 times faster than the vanilla diffusion policy and 10 times faster than the state-of-the-art MPD method, while matching the task success of multi-step diffusion policies. By targeting the diffusion component used throughout VLA systems, FRMD provides a plug-in, latency-optimized motion generator that preserves the advantages of diffusion and makes real-time embodied AI feasible.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。