Plotting

 Mesbahi, Mehran


Estimation-Aware Trajectory Optimization with Set-Valued Measurement Uncertainties

arXiv.org Artificial Intelligence

In this paper, we present an optimization-based framework for generating estimation-aware trajectories in scenarios where measurement (output) uncertainties are state-dependent and set-valued. The framework leverages the concept of regularity for set-valued output maps. Specifically, we demonstrate that, for output-regular maps, one can utilize a set-valued observability measure that is concave with respect to finite-horizon state trajectories. By maximizing this measure, optimized estimation-aware trajectories can be designed for a broad class of systems, including those with locally linearized dynamics. To illustrate the effectiveness of the proposed approach, we provide a representative example in the context of trajectory planning for vision-based estimation. We present an estimation-aware trajectory for an uncooperative target-tracking problem that uses a machine learning (ML)-based estimation module on an ego-satellite.


Multi Agent Reinforcement Learning for Sequential Satellite Assignment Problems

arXiv.org Artificial Intelligence

Assignment problems are a classic combinatorial optimization problem in which a group of agents must be assigned to a group of tasks such that maximum utility is achieved while satisfying assignment constraints. Given the utility of each agent completing each task, polynomial-time algorithms exist to solve a single assignment problem in its simplest form. However, in many modern-day applications such as satellite constellations, power grids, and mobile robot scheduling, assignment problems unfold over time, with the utility for a given assignment depending heavily on the state of the system. We apply multi-agent reinforcement learning to this problem, learning the value of assignments by bootstrapping from a known polynomial-time greedy solver and then learning from further experience. We then choose assignments using a distributed optimal assignment mechanism rather than by selecting them directly. We demonstrate that this algorithm is theoretically justified and avoids pitfalls experienced by other RL algorithms in this setting. Finally, we show that our algorithm significantly outperforms other methods in the literature, even while scaling to realistic scenarios with hundreds of agents and tasks.


Data-Guided Regulator for Adaptive Nonlinear Control

arXiv.org Artificial Intelligence

A critical aspect of autonomous operations in safety-critical scenarios is learning from available data for quick adaptation to new environments while maintaining safety. Examples include aircraft emergency landing scenarios in adverse weather conditions and agile quadrotor flights through low clearance gates in the presence of dynamic and strong wind conditions [1]. From a system theoretic perspective, this system feature maps to having the autonomous agent handle parametric model uncertainties and disturbances with control-theoretic guarantees such as stability and tracking error convergence, common in adaptive control settings [2, 3]. A rich body of literature has analyzed classical adaptive control algorithms' stability and convergence properties for continuous-time dynamical systems. Such studies include the use of PI (proportional integral) controllers [4] for a class of linear time-varying systems to guarantee (I) infinite-time convergence of the tracking error to zero, i.e., the difference between actual and nominal states () = () (), for any constant exogenous disturbance (denoted by), (II) infinite-time convergence of the tracking error () to a bound which is proportional to the bound on the magnitude of the rate of the exogenous signal ().


Data-Driven Structured Policy Iteration for Homogeneous Distributed Systems

arXiv.org Artificial Intelligence

Control of networked systems, comprised of interacting agents, is often achieved through modeling the underlying interactions. Constructing accurate models of such interactions--in the meantime--can become prohibitive in applications. Data-driven control methods avoid such complications by directly synthesizing a controller from the observed data. In this paper, we propose an algorithm referred to as Data-driven Structured Policy Iteration (D2SPI), for synthesizing an efficient feedback mechanism that respects the sparsity pattern induced by the underlying interaction network. In particular, our algorithm uses temporary "auxiliary" communication links in order to enable the required information exchange on a (smaller) sub-network during the "learning phase" -- links that will be removed subsequently for the final distributed feedback synthesis. We then proceed to show that the learned policy results in a stabilizing structured policy for the entire network. Our analysis is then followed by showing the stability and convergence of the proposed distributed policies throughout the learning phase, exploiting a construct referred to as the "Patterned monoid.'' The performance of D2SPI is then demonstrated using representative simulation scenarios.


Adaptive Traffic Control with Deep Reinforcement Learning: Towards State-of-the-art and Beyond

arXiv.org Machine Learning

In this work, we study adaptive data-guided traffic planning and control using Reinforcement Learning (RL). We shift from the plain use of classic methods towards state-of-the-art in deep RL community. We embed several recent techniques in our algorithm that improve the original Deep Q-Networks (DQN) for discrete control and discuss the traffic-related interpretations that follow. We propose a novel DQN-based algorithm for Traffic Control (called TC-DQN+) as a tool for fast and more reliable traffic decision-making. We introduce a new form of reward function which is further discussed using illustrative examples with comparisons to traditional traffic control methods.


Global Convergence of Policy Gradient Methods for Linearized Control Problems

arXiv.org Machine Learning

Direct policy gradient methods for reinforcement learning and continuous control problems are a popular approach for a variety of reasons: 1) they are easy to implement without explicit knowledge of the underlying model 2) they are an "end-to-end" approach, directly optimizing the performance metric of interest 3) they inherently allow for richly parameterized policies. A notable drawback is that even in the most basic continuous control problem (that of linear quadratic regulators), these methods must solve a non-convex optimization problem, where little is understood about their efficiency from both computational and statistical perspectives. In contrast, system identification and model based planning in optimal control theory have a much more solid theoretical footing, where much is known with regards to their computational and statistical properties. This work bridges this gap showing that (model free) policy gradient methods globally converge to the optimal solution and are efficient (polynomially so in relevant problem dependent quantities) with regards to their sample and computational complexities.