KDPE: A Kernel Density Estimation Strategy for Diffusion Policy Trajectory Selection

Abstract

We propose KDPE, a Kernel Density Estimation-based strategy that filters out potentially harmful trajectories output of Diffusion Policy while keeping a low test-time computational overhead. For Kernel Density Estimation, we propose a manifold-aware kernel to model a probability density function for actions composed of end-effector Cartesian position, orientation, and gripper state.

KDPE overall achieves better performance than Diffusion Policy on simulated single-arm RoboMimic and MimicGen tasks, and on three real robot experiments: PickPlush, a tabletop grasping task, CubeSort, a multimodal pick and place task, and CoffeeMaking, a task that requires long-horizon capabilities and precise execution.

Simulation Experiments

Benchmark

RoboMimic

Lift task animation

Lift

Can task animation

Can

Square task animation

Square

ToolHang task animation

ToolHang

MimicGen

Coffee task animation

Coffee

Stack task animation

Stack

Assembly task animation

Assembly

Real-world Experiments

Benchmark

KDPE

PickPlush
(success rate: 96%)

PickSponge
(success rate: 90%)

CubeSort
(success rate: 44%)

CoffeeMaking
(success rate: 70%)

Baseline

PickPlush
(success rate: 90%)

PickSponge
(success rate: 88%)

CubeSort
(success rate: 41%)

CoffeeMaking
(success rate: 60%)

Same Initialization

KDPE

Baseline

Trajectory Visualizer

Our trajectory visualizer is tool designed to analyze the distribution of trajectories sampled by generative robotic policies. It helps in understanding the policy's behavior by revealing patterns in the sampled trajectories, identifying potential failure cases, and quantifying the amount of outliers. By visualizing hundreds of trajectories simultaneously, we can better assess the policy's consistency and robustness across different scenarios.

Visualization of diffusion policy completing the ToolHang task.
Note: The color of the trajectories represents the gripper state at the last trajectory step.