Plotting

 Zhou, Chengxu


Learning to Adapt: Bio-Inspired Gait Strategies for Versatile Quadruped Locomotion

arXiv.org Artificial Intelligence

Deep reinforcement learning (DRL) has revolutionised quadruped robot locomotion, but existing control frameworks struggle to generalise beyond their training-induced observational scope, resulting in limited adaptability. In contrast, animals achieve exceptional adaptability through gait transition strategies, diverse gait utilisation, and seamless adjustment to immediate environmental demands. Inspired by these capabilities, we present a novel DRL framework that incorporates key attributes of animal locomotion: gait transition strategies, pseudo gait procedural memory, and adaptive motion adjustments. This approach enables our framework to achieve unparalleled adaptability, demonstrated through blind zero-shot deployment on complex terrains and recovery from critically unstable states. Our findings offer valuable insights into the biomechanics of animal locomotion, paving the way for robust, adaptable robotic systems.


Learning Bipedal Walking on a Quadruped Robot via Adversarial Motion Priors

arXiv.org Artificial Intelligence

Previous studies have successfully demonstrated agile and robust locomotion in challenging terrains for quadrupedal robots. However, the bipedal locomotion mode for quadruped robots remains unverified. This paper explores the adaptation of a learning framework originally designed for quadrupedal robots to operate blind locomotion in biped mode. We leverage a framework that incorporates Adversarial Motion Priors with a teacher-student policy to enable imitation of a reference trajectory and navigation on tough terrain. Our work involves transferring and evaluating a similar learning framework on a quadruped robot in biped mode, aiming to achieve stable walking on both flat and complicated terrains. Our simulation results demonstrate that the trained policy enables the quadruped robot to navigate both flat and challenging terrains, including stairs and uneven surfaces.


Deep Reinforcement Learning for Bipedal Locomotion: A Brief Survey

arXiv.org Artificial Intelligence

Bipedal robots are garnering increasing global attention due to their potential applications and advancements in artificial intelligence, particularly in Deep Reinforcement Learning (DRL). While DRL has driven significant progress in bipedal locomotion, developing a comprehensive and unified framework capable of adeptly performing a wide range of tasks remains a challenge. This survey systematically categorizes, compares, and summarizes existing DRL frameworks for bipedal locomotion, organizing them into end-to-end and hierarchical control schemes. End-to-end frameworks are assessed based on their learning approaches, whereas hierarchical frameworks are dissected into layers that utilize either learning-based methods or traditional model-based approaches. This survey provides a detailed analysis of the composition, capabilities, strengths, and limitations of each framework type. Furthermore, we identify critical research gaps and propose future directions aimed at achieving a more integrated and efficient framework for bipedal locomotion, with potential broad applications in everyday life.


Web-based Experiment on Human Performance in Dual-Robot Teleoperation

arXiv.org Artificial Intelligence

In most cases, upgrading from a single-robot system to a multi-robot system comes with increases in system payload and task performance. On the other hand, many multi-robot systems in open environments still rely on teleoperation. Therefore, human performance can be the bottleneck in a teleoperated multi-robot system. Based on this idea, the multi-robot system's shared autonomy and control methods are emerging research areas in open environment robot operations. However, the question remains: how much does the bottleneck of the human agent impact the system performance in a multi-robot system? This research tries to explore the question through the performance comparison of teleoperating a single-robot system and a dual-robot system in a box-pushing task. This robot teleoperation experiment on human agents employs a web-based environment to simulate the robots' two-dimensional movement. The result provides evidence of the hardship for a single human when teleoperating with more than one robot, which indicates the necessity of shared autonomy in multi-robot systems.


Nonlinear Model Predictive Control for Robust Bipedal Locomotion: Exploring Angular Momentum and CoM Height Changes

arXiv.org Artificial Intelligence

-- Human beings can utilize multiple balance strategies, e.g. In this work, we propose a novel Nonlinear Model Predictive Control (NMPC) framework for robust locomotion, with the capabilities of step location adjustment, Center of Mass (CoM) height variation, and angular momentum adaptation. These features are realized by constraining the Zero Moment Point within the support polygon. By using the nonlinear inverted pendulum plus flywheel model, the effects of upper-body rotation and vertical height motion are considered. As a result, the NMPC is formulated as a quadratically constrained quadratic program problem, which is solved fast by sequential quadratic programming. Using this unified framework, robust walking patterns that exploit reactive stepping, body inclination, and CoM height variation are generated based on the state estimation. The adaptability for bipedal walking in multiple scenarios has been demonstrated through simulation studies. Humanoid robots have attracted much attention for their capabilities in accomplishing challenging tasks in real-world environments. With several decades passed, state-of-the-art robot platforms such as ASIMO [1], Atlas [2], W ALK-MAN [3], and CogIMon [4] have been developed for this purpose. However, due to the complex nonlinear dynamics of bipedal locomotion over the walking process, enhancing walking stability, which is among the prerequisites in making humanoids practical, still needs further studies. In this paper, inspired by the fact that human beings can make use of the redundant Degree of Freedom (DoF) and adopt various strategies, such as the ankle, hip, and stepping strategies, to realize balance recovery [5]-[7], we aim to develop a versatile and robust walking pattern generator which can integrate multiple balance strategies in a unified way. To generate the walking pattern in a time-efficient manner, simplified dynamic models have been proposed, among which the Linear Inverted Pendulum Model (LIPM) is widely used [8]. Using the LIPM, Kajita et al. proposed the preview control for Zero Moment Point (ZMP) tracking [9]. By adopting a Linear Quadratic Regulator (LQR) scheme, the ankle torque was adjusted to modulate the ZMP trajectory and Center of Mass (CoM) trajectory. Nevertheless, this strategy can neither modulate the step parameters nor take into consideration the feasibility constraints arisen from actuation limitations and environmental constraints. To overcome this drawback, Wieber et al. proposed a Model Predictive Control (MPC) algorithm to utilize the ankle strategy [10] and then extended it for adjusting step location [11].