View a PDF of the paper titled Score and Distribution Matching Policy: Advanced Accelerated Visuomotor Policies via Matched Distillation, by Bofang Jia and 7 other authors
Abstract:Visual-motor policy learning has advanced with architectures like diffusion-based policies, known for modeling complex robotic trajectories. However, their prolonged inference times hinder high-frequency control tasks requiring real-time feedback. While consistency distillation (CD) accelerates inference, it introduces errors that compromise action quality. To address these limitations, we propose the Score and Distribution Matching Policy (SDM Policy), which transforms diffusion-based policies into single-step generators through a two-stage optimization process: score matching ensures alignment with true action distributions, and distribution matching minimizes KL divergence for consistency. A dual-teacher mechanism integrates a frozen teacher for stability and an unfrozen teacher for adversarial training, enhancing robustness and alignment with target distributions. Evaluated on a 57-task simulation benchmark, SDM Policy achieves a 6x inference speedup while having state-of-the-art action quality, providing an efficient and reliable framework for high-frequency robotic tasks.
Submission history
From: Bofang Jia [view email]
[v1]
Thu, 12 Dec 2024 13:22:02 UTC (2,207 KB)
[v2]
Fri, 13 Dec 2024 08:44:16 UTC (2,207 KB)
[v3]
Mon, 16 Dec 2024 05:43:20 UTC (2,210 KB)
Source link
lol