Coltin, Brian
Semantic Masking and Visual Feature Matching for Robust Localization
Mao, Luisa, Soussan, Ryan, Coltin, Brian, Smith, Trey, Biswas, Joydeep
We are interested in long-term deployments of autonomous robots to aid astronauts with maintenance and monitoring operations in settings such as the International Space Station. Unfortunately, such environments tend to be highly dynamic and unstructured, and their frequent reconfiguration poses a challenge for robust long-term localization of robots. Many state-of-the-art visual feature-based localization algorithms are not robust towards spatial scene changes, and SLAM algorithms, while promising, cannot run within the low-compute budget available to space robots. To address this gap, we present a computationally efficient semantic masking approach for visual feature matching that improves the accuracy and robustness of visual localization systems during long-term deployment in changing environments. Our method introduces a lightweight check that enforces matches to be within long-term static objects and have consistent semantic classes. We evaluate this approach using both map-based relocalization and relative pose estimation and show that it improves Absolute Trajectory Error (ATE) and correct match ratios on the publicly available Astrobee dataset. While this approach was originally developed for microgravity robotic freeflyers, it can be applied to any visual feature matching pipeline to improve robustness.
Multi-Agent 3D Map Reconstruction and Change Detection in Microgravity with Free-Flying Robots
Dinkel, Holly, Di, Julia, Santos, Jamie, Albee, Keenan, Borges, Paulo, Moreira, Marina, Alexandrov, Oleg, Coltin, Brian, Smith, Trey
Assistive free-flyer robots autonomously caring for future crewed outposts -- such as NASA's Astrobee robots on the International Space Station (ISS) -- must be able to detect day-to-day interior changes to track inventory, detect and diagnose faults, and monitor the outpost status. This work presents a framework for multi-agent cooperative mapping and change detection to enable robotic maintenance of space outposts. One agent is used to reconstruct a 3D model of the environment from sequences of images and corresponding depth information. Another agent is used to periodically scan the environment for inconsistencies against the 3D model. Change detection is validated after completing the surveys using real image and pose data collected by Astrobee robots in a ground testing environment and from microgravity aboard the ISS. This work outlines the objectives, requirements, and algorithmic modules for the multi-agent reconstruction system, including recommendations for its use by assistive free-flyers aboard future microgravity outposts. *Denotes Equal Contribution
Unsupervised Change Detection for Space Habitats Using 3D Point Clouds
Santos, Jamie, Dinkel, Holly, Di, Julia, Borges, Paulo V. K., Moreira, Marina, Alexandrov, Oleg, Coltin, Brian, Smith, Trey
This work presents an algorithm for scene change detection from point clouds to enable autonomous robotic caretaking in future space habitats. Autonomous robotic systems will help maintain future deep-space habitats, such as the Gateway space station, which will be uncrewed for extended periods. Existing scene analysis software used on the International Space Station (ISS) relies on manually-labeled images for detecting changes. In contrast, the algorithm presented in this work uses raw, unlabeled point clouds as inputs. The algorithm first applies modified Expectation-Maximization Gaussian Mixture Model (GMM) clustering to two input point clouds. It then performs change detection by comparing the GMMs using the Earth Mover's Distance. The algorithm is validated quantitatively and qualitatively using a test dataset collected by an Astrobee robot in the NASA Ames Granite Lab comprising single frame depth images taken directly by Astrobee and full-scene reconstructed maps built with RGB-D and pose data from Astrobee. The runtimes of the approach are also analyzed in depth. The source code is publicly released to promote further development.
An Investigation of Multi-feature Extraction and Super-resolution with Fast Microphone Arrays
Chang, Eric T., Wang, Runsheng, Ballentine, Peter, Xu, Jingxi, Smith, Trey, Coltin, Brian, Kymissis, Ioannis, Ciocarlie, Matei
In this work, we use MEMS microphones as vibration sensors to simultaneously classify texture and estimate contact position and velocity. Vibration sensors are an important facet of both human and robotic tactile sensing, providing fast detection of contact and onset of slip. Microphones are an attractive option for implementing vibration sensing as they offer a fast response and can be sampled quickly, are affordable, and occupy a very small footprint. Our prototype sensor uses only a sparse array of distributed MEMS microphones (8-9 mm spacing) embedded under an elastomer. We use transformer-based architectures for data analysis, taking advantage of the microphones' high sampling rate to run our models on time-series data as opposed to individual snapshots. This approach allows us to obtain 77.3% average accuracy on 4-class texture classification (84.2% when excluding the slowest drag velocity), 1.5 mm median error on contact localization, and 4.5 mm/s median error on contact velocity. We show that the learned texture and localization models are robust to varying velocity and generalize to unseen velocities. We also report that our sensor provides fast contact detection, an important advantage of fast transducers. This investigation illustrates the capabilities one can achieve with a MEMS microphone array alone, leaving valuable sensor real estate available for integration with complementary tactile sensing modalities.
CoBots: Robust Symbiotic Autonomous Mobile Service Robots
Veloso, Manuela (Carnegie Mellon University) | Biswas, Joydeep (Carnegie Mellon University) | Coltin, Brian (Carnegie Mellon University) | Rosenthal, Stephanie (Carnegie Mellon University)
We research and develop autonomous mobile service robots as Collaborative Robots, i.e., CoBots. For the last three years, our four CoBots have autonomously navigated in our multi-floor office buildings for more than 1,000km, as the result of the integration of multiple perceptual, cognitive, and actuations representations and algorithms. In this paper, we identify a few core aspects of our CoBots underlying their robust functionality. The reliable mobility in the varying indoor environments comes from a novel episodic non-Markov localization. Service tasks requested by users are the input to a scheduler that can consider different types of constraints, including transfers among multiple robots. With symbiotic autonomy, the CoBots proactively seek external sources of help to fill-in for their inevitable occasional limitations. We present sampled results from a deployment and conclude with a brief review of other features of our service robots.
Scheduling for Transfers in Pickup and Delivery Problems with Very Large Neighborhood Search
Coltin, Brian (The Robotics Institute, Carnegie Mellon University) | Veloso, Manuela (Carnegie Mellon University)
In pickup and delivery problems (PDPs), vehicles pickup and deliver a set of items under various constraints. We address the PDP with Transfers (PDP-T), in which vehicles plan to transfer items between one another to form more efficient schedules. We introduce the Very Large Neighborhood Search with Transfers (VLNS-T) algorithm to form schedules for the PDP-T. Our approach allows multiple transfers for items at arbitrary locations, and is not restricted to a set of predefined transfer points. We show that VLNS-T improves upon the best known PDP solutions for benchmark problems, and demonstrate its effectiveness on problems sampled from real world taxi data in New York City.
Web-Based Remote Assistance to Overcome Robot Perceptual Limitations
Ventura, Rodrigo (Universidade Técnica de Lisboa) | Coltin, Brian (Carnegie Mellon University) | Veloso, Manuela (Carnegie Mellon University)
This paper addresses the problem of overcoming visual perception limitations in service robots with remote assistance from human users. In particular, consider a scenario where a user requests the robot to perform some task that requires a perceptual ability, e.g., check if a specific mug, "my mug," is in the lab or in an office, but the robot may not know how to recognize that object. We propose to equip the robots with the abilities to: (i) identify their own perceptual limitations, (ii) autonomously and remotely query human users for assistance, and (iii) learn new object descriptors from the interaction with humans. We successfully developed a complete initial version of our approach on our CoBot service mobile robot. The interaction with the user builds upon our previously developed semi-autonomous telepresence image sharing and control. The user can now further identify the object and the robot can save the descriptor and use it in future situations. We illustrate our work with the task of learning to identify an object in the environment, and to report its presence to a user. Our ongoing work includes addressing a dynamic interaction between the robot and the remote user for visual focus of attention and different object viewing, as well as the effective storage, labeling, accessing, and sharing of multiple learned object descriptors, in particular among robots. Our goal is also to contribute the learned knowledge to crowd-robotics efforts.
Interruptable Autonomy: Towards Dialog-Based Robot Task Management
Sun, Yichao (Zhejiang University) | Coltin, Brian (Carnegie Mellon University) | Veloso, Manuela (Carnegie Mellon University)
We have been successfully deploying mobile service robots in an office building to execute user-requested tasks, such as delivering messages, transporting items, escorting people, and enabling telepresence. Users submit task requests to a dedicated website which forms a schedule of tasks for the robots to execute. The robots autonomously navigate in the building to complete their tasks. However, upon observing the many successful task executions, we realized that the robots are too autonomous in their determination to execute a planned task, with no mechanism to interrupt or redirect the robot through local interaction. In this work, we analyze the challenges of this goal of interruption, and contribute an approach to interrupt the robot anywhere during its execution through spoken dialog. Tasks can then be modified or new tasks can be added through speech, allowing users to manage the robot’s schedule. We discuss the response of the robot to human interruptions. We also introduce a finite state machine based on spoken dialog to handle the conversations that might occur during task execution. The goal is for the robot to fulfill humans’ requests as much as possible while minimizing the impact to the ongoing and pending tasks. We present examples of our task interruption scheme executing on robots to demonstrate its effectiveness.
Multi-Observation Sensor Resetting Localization with Ambiguous Landmarks
Coltin, Brian (Carnegie Mellon University) | Veloso, Manuela (Carnegie Mellon University)
Successful approaches to the robot localization problem include Monte Carlo particle filters, which estimate non-parametric localization belief distributions. However, particle filters fare poorly at determining the robot's position without a good initial hypothesis. This problem has been addressed for robots that sense visual landmarks with sensor resetting, by performing sensor-based resampling when the robot is lost. For robots that make sparse, ambiguous and noisy observations, standard sensor resetting places new location hypotheses across a wide region, in positions that may be inconsistent with previous observations. We propose Multi-Observation Sensor Resetting, where observations from multiple frames are merged to generate new hypotheses more effectively. We demonstrate experimentally in the robot soccer domain on the NAO humanoid robots that Multi-Observation Sensor Resetting converges more efficiently to the robot's true position than standard sensor resetting, and is more robust to systematic vision errors.