Goto

Collaborating Authors

 Siegwart, Roland


Safe Periodic Trochoidal Paths for Fixed-Wing UAVs in Confined Windy Environments

arXiv.org Artificial Intelligence

Safe Periodic Trochoidal Paths for Fixed-Wing UA Vs in Confined Windy Environments Jaeyoung Lim 1, David Rohr 1, Thomas Stastny 1, Roland Siegwart 1 Abstract -- Due to their energy-efficient flight characteristics, fixed-wing type uncrewed aerial vehicles (UA Vs) are useful robotic tools for long-range and duration flight applications in large-scale environments. However, flying fixed-wing UA V in confined environments, such as mountainous regions, can be challenging due to their limited maneuverability and sensitivity to uncertain wind conditions. In this work, we first analyze periodic trochoidal paths that can be used to define wind-aware terminal loitering states. We then propose a wind-invariant safe set of trochoidal paths along with a switching strategy for selecting the corresponding minimum-extent periodic path type. Finally, we show that planning with this minimum-extent set allows us to safely reach up to 10 times more locations in mountainous terrain compared to planning with a single, conservative loitering maneuver . I. INTRODUCTION Uncrewed aerial vehicles (UA Vs) have become crucial tools for information-gathering applications, such as surveying and inspection [1], search and rescue [2], and environment monitoring [3], [4]. For large-scale coverage or long-range applications, fixed-wing type UA Vs are preferred over rotary-wing type systems due to their high endurance and speed. While the wing-borne aerodynamic lift enables energy-efficient flight, it also poses challenges for operating safely.


Towards Open-Source and Modular Space Systems with ATMOS

arXiv.org Artificial Intelligence

Abstract--In the near future, autonomous space systems will compose a large number of the spacecraft being deployed. Their tasks will involve autonomous rendezvous and proximity operations with large structures, such as inspections or assembly of orbiting space stations and maintenance and human-assistance tasks over shared workspaces. To promote replicable and reliable scientific results for autonomous control of spacecraft, we present the design of a space systems laboratory based on open-source and modular software and hardware. The simulation software provides a software-in-the-loop (SITL) architecture that seamlessly transfers simulated results to the ATMOS platforms, developed for testing of multi-agent autonomy schemes for microgravity. The manuscript presents the KTH space systems laboratory facilities and the ATMOS platform as open-source hardware and software contributions. To the left, we see the tethers of the low-pressure compressor system. Software and hardware contributions can be found in: 1. PX4Space: Athens [6] proposed a similar test bed, where the platforms https://atmos.discower.io The facility also provides a vision-based I. This The space sector has experienced significant growth [1] in facility was recently upgraded to more modern avionics, motion the last decade, in part due to the decreased costs of access to capture ground-truth positioning, and robotics communication space through multiple commercial operators [2], but also due software through the Robotics Operating System (ROS) [7]. to the maturation of existing technologies and, consequently, Stanford University's Autonomous Systems Laboratory freeflyer reduced pricing of the deployed equipment. In the last twenty testbed [8], [9], [10] uses a similar, round platform as to thirty years, a few academic and industrial research facilities a free-flying robotic system for path planning, docking and have been created to test space systems by replicating motion capturing of space systems, paired with an open-source Python in microgravity on Earth.


Learning Affordances from Interactive Exploration using an Object-level Map

arXiv.org Artificial Intelligence

Many robotic tasks in real-world environments require physical interactions with an object such as pick up or push. For successful interactions, the robot needs to know the object's affordances, which are defined as the potential actions the robot can perform with the object. In order to learn a robot-specific affordance predictor, we propose an interactive exploration pipeline which allows the robot to collect interaction experiences while exploring an unknown environment. We integrate an object-level map in the exploration pipeline such that the robot can identify different object instances and track objects across diverse viewpoints. This results in denser and more accurate affordance annotations compared to state-of-the-art methods, which do not incorporate a map. We show that our affordance exploration approach makes exploration more efficient and results in more accurate affordance prediction models compared to baseline methods.


Allocation for Omnidirectional Aerial Robots: Incorporating Power Dynamics

arXiv.org Artificial Intelligence

Tilt-rotor aerial robots are more dynamic and versatile than their fixed-rotor counterparts, since the thrust vector and body orientation are decoupled. However, the coordination of servomotors and propellers (the allocation problem) is not trivial, especially accounting for overactuation and actuator dynamics. We present and compare different methods of actuator allocation for tilt-rotor platforms, evaluating them on a real aerial robot performing dynamic trajectories. We extend the state-of-the-art geometric allocation into a differential allocation, which uses the platform's redundancy and does not suffer from singularities typical of the geometric solution. We expand it by incorporating actuator dynamics and introducing propeller limit curves. These improve the modeling of propeller limits, automatically balancing their usage and allowing the platform to selectively activate and deactivate propellers during flight. We show that actuator dynamics and limits make the tuning of the allocation not only easier, but also allow it to track more dynamic oscillating trajectories with angular velocities up to 4 rad/s, compared to 2.8 rad/s of geometric methods.


Evaluation of Human-Robot Interfaces based on 2D/3D Visual and Haptic Feedback for Aerial Manipulation

arXiv.org Artificial Intelligence

Most telemanipulation systems for aerial robots provide the operator with only 2D screen visual information. The lack of richer information about the robot's status and environment can limit human awareness and, in turn, task performance. While the pilot's experience can often compensate for this reduced flow of information, providing richer feedback is expected to reduce the cognitive workload and offer a more intuitive experience overall. This work aims to understand the significance of providing additional pieces of information during aerial telemanipulation, namely (i) 3D immersive visual feedback about the robot's surroundings through mixed reality (MR) and (ii) 3D haptic feedback about the robot interaction with the environment. To do so, we developed a human-robot interface able to provide this information. First, we demonstrate its potential in a real-world manipulation task requiring sub-centimeter-level accuracy. Then, we evaluate the individual effect of MR vision and haptic feedback on both dexterity and workload through a human subjects study involving a virtual block transportation task. Results show that both 3D MR vision and haptic feedback improve the operator's dexterity in the considered teleoperated aerial interaction tasks. Nevertheless, pilot experience remains the most significant factor.


Obstacle-Avoidant Leader Following with a Quadruped Robot

arXiv.org Artificial Intelligence

Personal mobile robotic assistants are expected to find wide applications in industry and healthcare. For example, people with limited mobility can benefit from robots helping with daily tasks, or construction workers can have robots perform precision monitoring tasks on-site. However, manually steering a robot while in motion requires significant concentration from the operator, especially in tight or crowded spaces. This reduces walking speed, and the constant need for vigilance increases fatigue and, thus, the risk of accidents. This work presents a virtual leash with which a robot can naturally follow an operator. We use a sensor fusion based on a custom-built RF transponder, RGB cameras, and a LiDAR. In addition, we customize a local avoidance planner for legged platforms, which enables us to navigate dynamic and narrow environments. We successfully validate on the ANYmal platform the robustness and performance of our entire pipeline in real-world experiments.


Radar Meets Vision: Robustifying Monocular Metric Depth Prediction for Mobile Robotics

arXiv.org Artificial Intelligence

Mobile robots require accurate and robust depth measurements to understand and interact with the environment. While existing sensing modalities address this problem to some extent, recent research on monocular depth estimation has leveraged the information richness, yet low cost and simplicity of monocular cameras. These works have shown significant generalization capabilities, mainly in automotive and indoor settings. However, robots often operate in environments with limited scale cues, self-similar appearances, and low texture. In this work, we encode measurements from a low-cost mmWave radar into the input space of a state-of-the-art monocular depth estimation model. Despite the radar's extreme point cloud sparsity, our method demonstrates generalization and robustness across industrial and outdoor experiments. Our approach reduces the absolute relative error of depth predictions by 9-64% across a range of unseen, real-world validation datasets. Importantly, we maintain consistency of all performance metrics across all experiments and scene depths where current vision-only approaches fail. We further address the present deficit of training data in mobile robotics environments by introducing a novel methodology for synthesizing rendered, realistic learning datasets based on photogrammetric data that simulate the radar sensor observations for training. Our code, datasets, and pre-trained networks are made available at https://github.com/ethz-asl/radarmeetsvision.


Framework for Robust Localization of UUVs and Mapping of Net Pens

arXiv.org Artificial Intelligence

This paper presents a general framework integrating vision and acoustic sensor data to enhance localization and mapping in highly dynamic and complex underwater environments, with a particular focus on fish farming. The proposed pipeline is suited to obtain both the net-relative pose estimates of an Unmanned Underwater Vehicle (UUV) and the depth map of the net pen purely based on vision data. Furthermore, this paper presents a method to estimate the global pose of an UUV fusing the net-relative pose estimates with acoustic data. The pipeline proposed in this paper showcases results on datasets obtained from industrial-scale fish farms and successfully demonstrates that the vision-based TRU-Depth model, when provided with sparse depth priors from the FFT method and combined with the Wavemap method, can estimate both net-relative and global position of the UUV in real time and generate detailed 3D maps suitable for autonomous navigation and inspection purposes.


Pushing the Limits of Reactive Planning: Learning to Escape Local Minima

arXiv.org Artificial Intelligence

When does a robot planner need a map? Reactive methods that use only the robot's current sensor data and local information are fast and flexible, but prone to getting stuck in local minima. Is there a middle-ground between fully reactive methods and map-based path planners? In this paper, we investigate feed forward and recurrent networks to augment a purely reactive sensor-based planner, which should give the robot geometric intuition about how to escape local minima. We train on a large number of extremely cluttered worlds auto-generated from primitive shapes, and show that our system zero-shot transfers to real 3D man-made environments, and can handle up to 30% sensor noise without degeneration of performance. We also offer a discussion of what role network memory plays in our final system, and what insights can be drawn about the nature of reactive vs. map-based navigation.


NeuSurfEmb: A Complete Pipeline for Dense Correspondence-based 6D Object Pose Estimation without CAD Models

arXiv.org Artificial Intelligence

State-of-the-art approaches for 6D object pose estimation assume the availability of CAD models and require the user to manually set up physically-based rendering (PBR) pipelines for synthetic training data generation. Both factors limit the application of these methods in real-world scenarios. In this work, we present a pipeline that does not require CAD models and allows training a state-of-the-art pose estimator requiring only a small set of real images as input. Our method is based on a NeuS2 object representation, that we learn through a semi-automated procedure based on Structure-from-Motion (SfM) and object-agnostic segmentation. We exploit the novel-view synthesis ability of NeuS2 and simple cut-and-paste augmentation to automatically generate photorealistic object renderings, which we use to train the correspondence-based SurfEmb pose estimator. We evaluate our method on the LINEMOD-Occlusion dataset, extensively studying the impact of its individual components and showing competitive performance with respect to approaches based on CAD models and PBR data. We additionally demonstrate the ease of use and effectiveness of our pipeline on self-collected real-world objects, showing that our method outperforms state-of-the-art CAD-model-free approaches, with better accuracy and robustness to mild occlusions. To allow the robotics community to benefit from this system, we will publicly release it at https://www.github.com/ethz-asl/neusurfemb.