Goto

Collaborating Authors

 Mainprice, Jim


GraspME -- Grasp Manifold Estimator

arXiv.org Artificial Intelligence

In this paper, we introduce a Grasp Manifold Estimator (GraspME) to detect grasp affordances for objects directly in 2D camera images. To perform manipulation tasks autonomously it is crucial for robots to have such graspability models of the surrounding objects. Grasp manifolds have the advantage of providing continuously infinitely many grasps, which is not the case when using other grasp representations such as predefined grasp points. For instance, this property can be leveraged in motion optimization to define goal sets as implicit surface constraints in the robot configuration space. In this work, we restrict ourselves to the case of estimating possible end-effector positions directly from 2D camera images. To this extend, we define grasp manifolds via a set of key points and locate them in images using a Mask R-CNN backbone. Using learned features allows generalizing to different view angles, with potentially noisy images, and objects that were not part of the training set. We rely on simulation data only and perform experiments on simple and complex objects, including unseen ones. Our framework achieves an inference speed of 11.5 fps on a GPU, an average precision for keypoint estimation of 94.5% and a mean pixel distance of only 1.29. This shows that we can estimate the objects very well via bounding boxes and segmentation masks as well as approximate the correct grasp manifold's keypoint coordinates.


A System for Traded Control Teleoperation of Manipulation Tasks using Intent Prediction from Hand Gestures

arXiv.org Artificial Intelligence

This paper presents a teleoperation system that includes robot perception and intent prediction from hand gestures. The perception module identifies the objects present in the robot workspace and the intent prediction module which object the user likely wants to grasp. This architecture allows the approach to rely on traded control instead of direct control: we use hand gestures to specify the goal objects for a sequential manipulation task, the robot then autonomously generates a grasping or a retrieving motion using trajectory optimization. The perception module relies on the model-based tracker to precisely track the 6D pose of the objects and makes use of a state of the art learning-based object detection and segmentation method, to initialize the tracker by automatically detecting objects in the scene. Goal objects are identified from user hand gestures using a trained a multi-layer perceptron classifier. After presenting all the components of the system and their empirical evaluation, we present experimental results comparing our pipeline to a direct traded control approach (i.e., one that does not use prediction) which shows that using intent prediction allows to bring down the overall task execution time.


Learning Cost Functions for Motion Planning of Human-Robot Collaborative Manipulation Tasks from Human-Human Demonstration

AAAI Conferences

In this work we present a method that allows to learn a cost function for motion planning of human-robot collaborative manipulation tasks where the human and the robot manipulate objects simultaneously in close proximity. Our approach is based on inverse optimal control which enables, considering a set of demonstrations, to find a cost function balancing different features. The cost function that is recovered from the human demonstrations is composed of elementary features, which are designed to encode notions such as safely, legibility and efficiency of the manipulation motions. We demonstrate the approach on data gathered from motion capture of human-human manipulation in close proximity of blocks on a table. To demonstrate the feasibility and efficacy of our approach we provide initial test results consisting of learning a cost function and then planning for the human kinematic model used in the learning phase.