Goto

Collaborating Authors

 maneuver


MIT engineers design an aerial microrobot that can fly as fast as a bumblebee

Robohub

In the future, tiny flying robots could be deployed to aid in the search for survivors trapped beneath the rubble after a devastating earthquake. So far, aerial microrobots have only been able to fly slowly along smooth trajectories, far from the swift, agile flight of real insects -- until now. MIT researchers have demonstrated aerial microrobots that can fly with speed and agility that is comparable to their biological counterparts. A collaborative team designed a new AI-based controller for the robotic bug that enabled it to follow gymnastic flight paths, such as executing continuous body flips. With a two-part control scheme that combines high performance with computational efficiency, the robot's speed and acceleration increased by about 450 percent and 250 percent, respectively, compared to the researchers' best previous demonstrations.


A Markov Decision Process Framework for Early Maneuver Decisions in Satellite Collision Avoidance

Ferrara, Francesca, Arana, Lander W. Schillinger, Dörfler, Florian, Li, Sarah H. Q.

arXiv.org Artificial Intelligence

ABSTRACT We develop a Markov decision process (MDP) framework to autonomously make guidance decisions for satellite collision avoidance maneuver (CAM) and a reinforcement learning policy gradient (RL-PG) algorithm to enable direct optimization of guidance policy using historic CAM data. In addition to maintaining acceptable collision risks, this approach seeks to minimize the average propellant consumption of CAMs by making early maneuver decisions. We model CAM as a continuous state, discrete action and finite horizon MDP, where the critical decision is determining when to initiate the maneuver. By deciding to maneuver earlier than conventional methods, the Markov policy effectively favors CAMs that achieve comparable rates of collision risk reduction while consuming less propellant. Using historical data of tracked conjunction events, we verify this framework and conduct an extensive parameter-sensitivity study. When evaluated on synthetic conjunction events, the trained policy consumes significantly less propellant overall and per maneuver in comparison to a conventional cut-off policy that initiates maneuvers 24 hours before the time of closest approach (TCA). On historical conjunction events, the trained policy consumes more propellant overall but consumes less propellant per maneuver. For both historical and synthetic conjunction events, the trained policy is slightly more conservative in identifying conjunctions events that warrant CAMs in comparison to cutoff policies.


Flow-Aided Flight Through Dynamic Clutters From Point To Motion

Xu, Bowen, Yan, Zexuan, Lu, Minghao, Fan, Xiyu, Luo, Yi, Lin, Youshen, Chen, Zhiqiang, Chen, Yeke, Qiao, Qiyuan, Lu, Peng

arXiv.org Artificial Intelligence

Challenges in traversing dynamic clutters lie mainly in the efficient perception of the environmental dynamics and the generation of evasive behaviors considering obstacle movement. Previous solutions have made progress in explicitly modeling the dynamic obstacle motion for avoidance, but this key dependency of decision-making is time-consuming and unreliable in highly dynamic scenarios with occlusions. On the contrary, without introducing object detection, tracking, and prediction, we empower the reinforcement learning (RL) with single LiDAR sensing to realize an autonomous flight system directly from point to motion. For exteroception, a depth sensing distance map achieving fixed-shape, low-resolution, and detail-safe is encoded from raw point clouds, and an environment change sensing point flow is adopted as motion features extracted from multi-frame observations. These two are integrated into a lightweight and easy-to-learn representation of complex dynamic environments. For action generation, the behavior of avoiding dynamic threats in advance is implicitly driven by the proposed change-aware sensing representation, where the policy optimization is indicated by the relative motion modulated distance field. With the deployment-friendly sensing simulation and dynamics model-free acceleration control, the proposed system shows a superior success rate and adaptability to alternatives, and the policy derived from the simulator can drive a real-world quadrotor with safe maneuvers.


Characterizing Human Feedback-Based Control in Naturalistic Driving Interactions via Gaussian Process Regression with Linear Feedback

DiPirro, Rachel, Devonport, Rosalyn, Calderone, Dan, Yang, Chishang "Mario'', Ju, Wendy, Oishi, Meeko

arXiv.org Artificial Intelligence

Understanding driver interactions is critical to designing autonomous vehicles to interoperate safely with human-driven cars. We consider the impact of these interactions on the policies drivers employ when navigating unsigned intersections in a driving simulator. The simulator allows the collection of naturalistic decision-making and behavior data in a controlled environment. Using these data, we model the human driver responses as state-based feedback controllers learned via Gaussian Process regression methods. We compute the feedback gain of the controller using a weighted combination of linear and nonlinear priors. We then analyze how the individual gains are reflected in driver behavior. We also assess differences in these controllers across populations of drivers. Our work in data-driven analyses of how drivers determine their policies can facilitate future work in the design of socially responsive autonomy for vehicles.


From Real-World Traffic Data to Relevant Critical Scenarios

Lüttner, Florian, Neis, Nicole, Stadler, Daniel, Moss, Robin, Fehling-Kaschek, Mirjam, Pfriem, Matthias, Stolz, Alexander, Ziehn, Jens

arXiv.org Artificial Intelligence

The reliable operation of autonomous vehicles, automated driving functions, and advanced driver assistance systems across a wide range of relevant scenarios is critical for their development and deployment. Identifying a near-complete set of relevant driving scenarios for such functionalities is challenging due to numerous degrees of freedom involved, each affecting the outcomes of the driving scenario differently. Moreover, with increasing technical complexity of new functionalities, the number of potentially relevant, particularly "unknown unsafe" scenarios is increasing. To enhance validation efficiency, it is essential to identify relevant scenarios in advance, starting with simpler domains like highways before moving to more complex environments such as urban traffic. To address this, this paper focuses on analyzing lane change scenarios in highway traffic, which involve multiple degrees of freedom and present numerous safetyrelevant scenarios. We describe the process of data acquisition and processing of real-world data from public highway traffic, followed by the application of criticality measures on trajectory data to evaluate scenarios, as conducted within the AVEAS project (www.aveas.org). By linking the calculated measures to specific lane change driving scenarios and the conditions under which the data was collected, we facilitate the identification of safetyrelevant driving scenarios for various applications. Further, to tackle the extensive range of "unknown unsafe" scenarios, we propose a way to generate relevant scenarios by creating synthetic scenarios based on recorded ones. Consequently, we demonstrate and evaluate a processing chain that enables the identification of safety-relevant scenarios, the development of data-driven methods for extracting these scenarios, and the generation of synthetic critical scenarios via sampling on highways.


Real-time Remote Tracking and Autonomous Planning for Whale Rendezvous using Robots

Bhattacharya, Sushmita, Jadhav, Ninad, Izhar, Hammad, Li, Karen, George, Kevin, Wood, Robert, Gil, Stephanie

arXiv.org Artificial Intelligence

We introduce a system for real-time sperm whale rendezvous at sea using an autonomous uncrewed aerial vehicle. Our system employs model-based reinforcement learning that combines in situ sensor data with an empirical whale dive model to guide navigation decisions. Key challenges include (i) real-time acoustic tracking in the presence of multiple whales, (ii) distributed communication and decision-making for robot deployments, and (iii) on-board signal processing and long-range detection from fish-trackers. We evaluate our system by conducting rendezvous with sperm whales at sea in Dominica, performing hardware experiments on land, and running simulations using whale trajectories interpolated from marine biologists' surface observations.


Crossing the Sim2Real Gap Between Simulation and Ground Testing to Space Deployment of Autonomous Free-flyer Control

Stewart, Kenneth, Chapin, Samantha, Leontie, Roxana, Henshaw, Carl Glen

arXiv.org Artificial Intelligence

Abstract-- Reinforcement learning (RL) offers transforma-tive potential for robotic control in space. We present the first on-orbit demonstration of RL-based autonomous control of a free-flying robot, the NASA Astrobee, aboard the International Space Station (ISS). Using NVIDIA's Omniverse physics simulator and curriculum learning, we trained a deep neural network to replace Astrobee's standard attitude and translation control, enabling it to navigate in microgravity. This successful deployment demonstrates the feasibility of training RL policies terrestrially and transferring them to space-based applications. This paves the way for future work in In-Space Servicing, Assembly, and Manufacturing (ISAM), enabling rapid on-orbit adaptation to dynamic mission requirements. Future In-Space Servicing, Assembly, and Manufacturing (ISAM) missions require increasingly autonomous robotic systems capable of adapting to the dynamic and uncertain conditions of space.


Autonomous Planning In-space Assembly Reinforcement-learning free-flYer (APIARY) International Space Station Astrobee Testing

Chapin, Samantha, Stewart, Kenneth, Leontie, Roxana, Henshaw, Carl Glen

arXiv.org Artificial Intelligence

The US Naval Research Laboratory's (NRL's) Autonomous Planning In-space Assembly Reinforcement-learning free-flYer (APIARY) experiment pioneers the use of reinforcement learning (RL) for control of free-flying robots in the zero-gravity (zero-G) environment of space. On Tuesday, May 27th 2025 the APIARY team conducted the first ever, to our knowledge, RL control of a free-flyer in space using the NASA Astrobee robot on-board the International Space Station (ISS). A robust 6-degrees of freedom (DOF) control policy was trained using an actor-critic Proximal Policy Optimization (PPO) network within the NVIDIA Isaac Lab simulation environment, randomizing over goal poses and mass distributions to enhance robustness. This paper details the simulation testing, ground testing, and flight validation of this experiment. This on-orbit demonstration validates the transformative potential of RL for improving robotic autonomy, enabling rapid development and deployment (in minutes to hours) of tailored behaviors for space exploration, logistics, and real-time mission needs.


Autonomous Reinforcement Learning Robot Control with Intel's Loihi 2 Neuromorphic Hardware

Stewart, Kenneth, Leontie, Roxana, Chapin, Samantha, Hays, Joe, Shrestha, Sumit Bam, Henshaw, Carl Glen

arXiv.org Artificial Intelligence

Abstract-- W e present an end-to-end pipeline for deploying reinforcement learning (RL) trained Artificial Neural Networks (ANNs) on neuromorphic hardware by converting them into spiking Sigma-Delta Neural Networks (SDNNs). W e demonstrate that an ANN policy trained entirely in simulation can be transformed into an SDNN compatible with Intel's Loihi 2 architecture, enabling low-latency and energy-efficient inference. As a test case, we use an RL policy for controlling the Astrobee free-flying robot, similar to a previously hardware in space-validated controller. The policy, trained with Rectified Linear Units (ReLUs), is converted to an SDNN and deployed on Intel's Loihi 2, then evaluated in NVIDIA's Omniverse Isaac Lab simulation environment for closed-loop control of Astrobee's motion. W e compare execution performance between GPU and Loihi 2. The results highlight the feasibility of using neuromorphic platforms for robotic control and establish a pathway toward energy-efficient, real-time neuromorphic computation in future space and terrestrial robotics applications.


Threat-Aware UAV Dodging of Human-Thrown Projectiles with an RGB-D Camera

Zhang, Yuying, Fan, Na, Zheng, Haowen, Liang, Junning, Pan, Zongliang, Chen, Qifeng, Lyu, Ximin

arXiv.org Artificial Intelligence

HE rapid advancement of uncrewed aerial vehicles (UA Vs) and their supporting infrastructure has significantly expanded the UA V market, enabling diverse applications such as aerial imaging, last-mile delivery, and air traffic management [1], [2]. To meet the demands of these complex tasks, modern UA Vs are increasingly equipped with autonomous modules for environmental perception, navigation, and obstacle avoidance. Despite these advances, UA Vs often fail to cope with sudden human-initiated attacks. Recent reports have documented cases where crowds at public events throw projectiles to disrupt UA V operations [3], [4], posing significant threats to their safety and public security. Consequently, there is an urgent need for robust strategies to counter human-initiated attacks involving fast-moving projectiles. Developing robust UA V systems capable of rapid responses to sudden human-initiated attacks remains a critical and unresolved research problem. Dodging such projectile threats involves overcoming several challenges: (1) Perception Latency: Projectiles often emerge suddenly at close range, leaving a narrow time window for detection and dodging. Therefore, minimizing the delay between sensing and control is crucial while maintaining high prediction accuracy to ensure effective avoidance.