Sampling-based Exploration for Reinforcement Learning of Dexterous Manipulation
Khandate, Gagan, Shang, Siqi, Chang, Eric T., Saidi, Tristan Luca, Liu, Yang, Dennis, Seth Matthew, Adams, Johnson, Ciocarlie, Matei
–arXiv.org Artificial Intelligence
Abstract--In this paper, we present a novel method for achieving dexterous manipulation of complex objects, while simultaneously securing the object without the use of passive support surfaces. We posit that a key difficulty for training such policies in a Reinforcement Learning framework is the difficulty of exploring the problem state space, as the accessible regions of this space form a complex structure along manifolds of a high-dimensional space. To address this challenge, we use two versions of the non-holonomic Rapidly-Exploring Random Trees algorithm; one version is more general, but requires explicit use of the environment's transition function, while the second version uses manipulation-specific kinematic constraints to attain better sample efficiency. In both cases, we use states found via sampling-based exploration to generate reset distributions that enable training control policies under full dynamic constraints via model-free Reinforcement Learning. We show that these policies are effective at manipulation problems of higher difficulty than previously shown, and also transfer effectively to real robots. Figure 1: Our method enables finger-gaiting manipulation of concave A number of example videos can also be found on the project or elongated objects which require complex gaits. Reinforcement Learning (RL) of robot sensorimotor control policies has seen great advances in recent years, demonstrated and highly effective family of Sampling-Based Planning (SBP) for a wide range of motor tasks.
arXiv.org Artificial Intelligence
May-23-2023
- Country:
- North America > United States (0.46)
- Genre:
- Research Report > New Finding (0.68)
- Industry:
- Energy (0.46)
- Leisure & Entertainment > Games (0.46)
- Technology: