Policy learning in robot-assisted surgery (RAS) lacks data efficient and versatile methods that exhibit the desired motion quality for delicate surgical interventions. To this end, we introduce Movement Primitive Diffusion (MPD), a novel method for imitation learning in RAS that focuses on gentle manipulation of deformable objects. The approach combines the versatility of diffusion-based imitation learning (DIL) with the high-quality motion generation capabilities of Probabilistic Dynamic Movement Primitives. This combination enables MPD to achieve gentle manipulation of deformable objects, while maintaining data efficiency critical for RAS applications where demonstration data is scarce. We evaluate MPD across various simulated and real world robotic tasks on both state and image observations. MPD outperforms state-of-the-art DIL methods in success rate, motion quality, and data efficiency.

TL;DR: Movement Primitive Diffusion (MPD) is a diffusion-based imitation learning method for high-quality robotic motion generation that focuses on gentle manipulation of deformable objects.


IEEE Xplore | arXiv | Code | Data | Video


Simulation Tasks
Real-World Tasks
Bimanual Tissue Manipulation Grasp Lift Touch Rope Threading Ligating Loop
Indirect deformable object manipulation, and concurrent bimanual coordination. Grasping, direct deformable object manipulation, and sequential bimanual coordination. Precision navigation of surgical thread and spatial reasoning. Precision navigation of a deformable instrument to constrict a deformable object.

Motion Quality Metrics

Baselines: BESO | DP-C | DP-T


Architectural Overview

MPD iteratively refines a noisy sequence of actions into a denoised vector of weights for a Probabilistic Dynamic Movement Primitive (ProDMP). The final weight vector is decoded into a smooth, high-frequency motion trajectory with guaranteed initial values for position and velocity.


Multi-Modal Behaviors

MPD is able to represent multi-modal behaviors for the same task, generating versatile motions. The mode of the generated motion depends on the initial random sample that is refined by MPD.

BibTeX

@article{Scheikl2024MPD,
  author={Scheikl, Paul Maria and Schreiber, Nicolas and Haas, Christoph and Freymuth, Niklas and Neumann, Gerhard and Lioutikov, Rudolf and Mathis-Ullrich, Franziska},
  title={Movement Primitive Diffusion: Learning Gentle Robotic Manipulation of Deformable Objects},
  journal={IEEE Robotics and Automation Letters},
  year={2024},
  volume={9},
  number={6},
  pages={5338-5345},
  doi={10.1109/LRA.2024.3382529},
}

Acknowledgements

The work was supported by the Erlangen National High Performance Computing Center funded by the German Research Foundation (DFG), the HoreKa supercomputer funded by the Ministry of Science, Research and the Arts Baden-Württemberg and by the Federal Ministry of Education and Research, and the DFG – 448648559.