Abstract
Learning policies for contact-rich manipulation is a challenging problem due to the presence of multiple contact modes with different dynamics, which complicates state and action exploration. Contact-rich motion planning uses simplified dynamics to reduce the search space dimension, but the found plans are then difficult to execute under the true object-manipulator dynamics. This paper presents an algorithm for learning controllers based on guided policy search, where motion plans based on simplified dynamics define rewards and sampling distributions for policy gradient-based learning. We demonstrate that our guided policy search method improves the ability to learn manipulation controllers, through a task involving pushing a box over a step.
Original language | English |
---|---|
Pages (from-to) | 1019-1024 |
Number of pages | 6 |
Journal | IFAC Proceedings Volumes (IFAC-PapersOnline) |
Volume | 58 |
Issue number | 28 |
DOIs | |
State | Published - 1 Oct 2024 |
Event | 4th Modeling, Estimation, and Control Conference, MECC 2024 - Chicago, United States Duration: 27 Oct 2024 → 30 Oct 2024 |
Keywords
- artificial intelligence
- machine learning
- motion planning
- Robotics
- robust control