Au, K. W. Samuel
DOFS: A Real-world 3D Deformable Object Dataset with Full Spatial Information for Dynamics Model Learning
Zhang, Zhen, Chu, Xiangyu, Tang, Yunxi, Au, K. W. Samuel
Robot manipulation of 3D Deformable Objects is essential for many activities and applications in the real world, such as household [1, 2] and healthcare [3], and is still an open challenge despite extensive studies. Recently, data-driven solutions have shown impressive and promising results in 3D deformable object manipulation by learning-based approaches [4, 5], where sufficient data is essential to improve model training or policy learning. To obtain training data, some previous works collected synthetic data from simulators [6]. Still, there is an unavoidable gap between the real world and the simulator since the existing simulators cannot accurately simulate all real-world physical characteristics (e.g., friction, impact, and stiffness) [7]. To mitigate the gap, some researchers [8, 9, 10, 11] collect Real-World Data (RWD); for example, [8, 9] collects RGB-D images and point clouds, [10] collects 3D mesh models, [11] uses a professional system with 106 cameras to obtain the 3D reconstructions of deformed mesh.
Interactive Navigation in Environments with Traversable Obstacles Using Large Language and Vision-Language Models
Zhang, Zhen, Lin, Anran, Wong, Chun Wai, Chu, Xiangyu, Dou, Qi, Au, K. W. Samuel
This paper proposes an interactive navigation framework by using large language and vision-language models, allowing robots to navigate in environments with traversable obstacles. We utilize the large language model (GPT-3.5) and the open-set Vision-language Model (Grounding DINO) to create an action-aware costmap to perform effective path planning without fine-tuning. With the large models, we can achieve an end-to-end system from textual instructions like "Can you pass through the curtains to deliver medicines to me?", to bounding boxes (e.g., curtains) with action-aware attributes. They can be used to segment LiDAR point clouds into two parts: traversable and untraversable parts, and then an action-aware costmap is constructed for generating a feasible path. The pre-trained large models have great generalization ability and do not require additional annotated data for training, allowing fast deployment in the interactive navigation tasks. We choose to use multiple traversable objects such as curtains and grasses for verification by instructing the robot to traverse them. Besides, traversing curtains in a medical scenario was tested. All experimental results demonstrated the proposed framework's effectiveness and adaptability to diverse environments.
Model-Free Large-Scale Cloth Spreading With Mobile Manipulation: Initial Feasibility Study
Chu+, Xiangyu, Wang+, Shengzhi, Feng, Minjian, Zheng, Jiaxi, Zhao, Yuxuan, Huang, Jing, Au, K. W. Samuel
Cloth manipulation is common in domestic and service tasks, and most studies use fixed-base manipulators to manipulate objects whose sizes are relatively small with respect to the manipulators' workspace, such as towels, shirts, and rags. In contrast, manipulation of large-scale cloth, such as bed making and tablecloth spreading, poses additional challenges of reachability and manipulation control. To address them, this paper presents a novel framework to spread large-scale cloth, with a single-arm mobile manipulator that can solve the reachability issue, for an initial feasibility study. On the manipulation control side, without modeling highly deformable cloth, a vision-based manipulation control scheme is applied and based on an online-update Jacobian matrix mapping from selected feature points to the end-effector motion. To coordinate the control of the manipulator and mobile platform, Behavior Trees (BTs) are used because of their modularity. Finally, experiments are conducted, including validation of the model-free manipulation control for cloth spreading in different conditions and the large-scale cloth spreading framework. The experimental results demonstrate the large-scale cloth spreading task feasibility with a single-arm mobile manipulator and the model-free deformation controller.
Inequality Constrained Trajectory Optimization with A Hybrid Multiple-shooting iLQR
Tang, Yunxi, Chu, Xiangyu, Jin, Wanxin, Au, K. W. Samuel
Trajectory optimization has been used extensively in robotic systems. In particular, iterative Linear Quadratic Regulator (iLQR) has performed well as an off-line planner and online nonlinear model predictive control solver, with a lower computational cost. However, standard iLQR cannot handle any constraints or perform reasonable initialization of a state trajectory. In this paper, we propose a hybrid constrained iLQR variant with a multiple-shooting framework to incorporate general inequality constraints and infeasible states initialization. The main technical contributions are twofold: 1) In addition to inheriting the simplicity of the initialization in multiple-shooting settings, a two-stage framework is developed to deal with state and/or control constraints robustly without loss of the linear feedback term of iLQR. Such a hybrid strategy offers fast convergence of constraint satisfaction. 2) An improved globalization strategy is proposed to exploit the coupled effects between line-searching and regularization, which is able to enhance the numerical robustness of the constrained iLQR approaches. Our approach is tested on various constrained trajectory optimization problems and outperforms the commonly-used collocation and shooting methods.
Towards Safe Landing of Falling Quadruped Robots Using a 3-DoF Morphable Inertial Tail
Tang, Yunxi, An, Jiajun, Chu, Xiangyu, Wang, Shengzhi, Wong, Ching Yan, Au, K. W. Samuel
Falling cat problem is well-known where cats show their super aerial reorientation capability and can land safely. For their robotic counterparts, a similar falling quadruped robot problem, has not been fully addressed, although achieving safe landing as the cats has been increasingly investigated. Unlike imposing the burden on landing control, we approach to safe landing of falling quadruped robots by effective flight phase control. Different from existing work like swinging legs and attaching reaction wheels or simple tails, we propose to deploy a 3-DoF morphable inertial tail on a medium-size quadruped robot. In the flight phase, the tail with its maximum length can self-right the body orientation in 3D effectively; before touch-down, the tail length can be retracted to about 1/4 of its maximum for impressing the tail's side-effect on landing. To enable aerial reorientation for safe landing in the quadruped robots, we design a control architecture, which has been verified in a high-fidelity physics simulation environment with different initial conditions. Experimental results on a customized flight-phase test platform with comparable inertial properties are provided and show the tail's effectiveness on 3D body reorientation and its fast retractability before touch-down. An initial falling quadruped robot experiment is shown, where the robot Unitree A1 with the 3-DoF tail can land safely subject to non-negligible initial body angles.