Abstract
Learning policies for contact-rich manipulation is a challenging problem due to the presence of multiple contact modes with different dynamics, which complicates state and action exploration. Contact-rich motion planning uses simplified dynamics to reduce the search space dimension, but the found plans are then difficult to execute under the true object-manipulator dynamics. This paper presents an algorithm for learning controllers based on guided policy search, where motion plans based on simplified dynamics define rewards and sampling distributions for policy gradient-based learning. We demonstrate that our guided policy search method improves the ability to learn manipulation controllers, through a task involving pushing a box over a step.
| Original language | English |
|---|---|
| Pages (from-to) | 1019-1024 |
| Number of pages | 6 |
| Journal | IFAC Proceedings Volumes (IFAC-PapersOnline) |
| Volume | 58 |
| Issue number | 28 |
| DOIs | |
| State | Published - 1 Oct 2024 |
| Event | 4th Modeling, Estimation, and Control Conference, MECC 2024 - Chicago, United States Duration: 27 Oct 2024 → 30 Oct 2024 |
Keywords
- artificial intelligence
- machine learning
- motion planning
- Robotics
- robust control