Not enough data to create a plot.
Try a different view from the menu above.
Weng, Thomas
On-Robot Reinforcement Learning with Goal-Contrastive Rewards
Biza, Ondrej, Weng, Thomas, Sun, Lingfeng, Schmeckpeper, Karl, Kelestemur, Tarik, Ma, Yecheng Jason, Platt, Robert, van de Meent, Jan-Willem, Wong, Lawson L. S.
Reinforcement Learning (RL) has the potential to enable robots to learn from their own actions in the real world. Unfortunately, RL can be prohibitively expensive, in terms of on-robot runtime, due to inefficient exploration when learning from a sparse reward signal. Designing dense reward functions is labour-intensive and requires domain expertise. In our work, we propose GCR (Goal-Contrastive Rewards), a dense reward function learning method that can be trained on passive video demonstrations. By using videos without actions, our method is easier to scale, as we can use arbitrary videos. GCR combines two loss functions, an implicit value loss function that models how the reward increases when traversing a successful trajectory, and a goal-contrastive loss that discriminates between successful and failed trajectories. We perform experiments in simulated manipulation environments across RoboMimic and MimicGen tasks, as well as in the real world using a Franka arm and a Spot quadruped. We find that GCR leads to a more-sample efficient RL, enabling model-free RL to solve about twice as many tasks as our baseline reward learning methods. We also demonstrate positive cross-embodiment transfer from videos of people and of other robots performing a task. Appendix: \url{https://tinyurl.com/gcr-appendix-2}.
Neural Grasp Distance Fields for Robot Manipulation
Weng, Thomas, Held, David, Meier, Franziska, Mukadam, Mustafa
Abstract-- We formulate grasp learning as a neural field and present Neural Grasp Distance Fields (NGDF). Here, the input is a 6D pose of a robot end effector and output is a distance to a continuous manifold of valid grasps for an object. In contrast to current approaches that predict a set of discrete candidate grasps, the distance-based NGDF representation is easily interpreted as a cost, and minimizing this cost produces a successful grasp pose. This grasp distance cost can be incorporated directly into a trajectory optimizer for joint optimization with other costs such as trajectory smoothness and collision avoidance. Figure 1: (a) Existing grasp estimation methods produce discrete grasp We evaluate NGDF on joint grasp and motion planning in sets which do not represent the true continuous manifold of possible simulation and the real world, outperforming baselines by 63% grasps. This distance can be leveraged as a cost for optimization, facilitating joint grasp and motion planning. We present Neural Grasp Distance Fields (NGDF), which optimization results in a smooth, collision-free trajectory that model the continuous manifold of valid grasp poses as the reaches a valid grasp pose.