Hardware Conditioned Policies for Multi-Robot Transfer Learning
Chen, Tao, Murali, Adithyavairavan, Gupta, Abhinav
–Neural Information Processing Systems
Deep reinforcement learning could be used to learn dexterous robotic policies but it is challenging to transfer them to new robots with vastly different hardware properties. It is also prohibitively expensive to learn a new policy from scratch for each robot hardware due to the high sample complexity of modern state-of-the-art algorithms. We propose a novel approach called Hardware Conditioned Policies where we train a universal policy conditioned on a vector representation of robot hardware. We considered robots in simulation with varied dynamics, kinematic structure, kinematic lengths and degrees-of-freedom. First, we use the kinematic structure directly as the hardware encoding and show great zero-shot transfer to completely novel robots not seen during training.
Neural Information Processing Systems
Feb-14-2020, 20:42:25 GMT