Briden, Julia
Fine-Tuned Language Models as Space Systems Controllers
Zucchelli, Enrico M., Wu, Di, Briden, Julia, Hofmann, Christian, Rodriguez-Fernandez, Victor, Linares, Richard
Large language models (LLMs), or foundation models (FMs), are pretrained transformers that coherently complete sentences auto-regressively. In this paper, we show that LLMs can control simplified space systems after some additional training, called fine-tuning. We look at relatively small language models, ranging between 7 and 13 billion parameters. We focus on four problems: a three-dimensional spring toy problem, low-thrust orbit transfer, low-thrust cislunar control, and powered descent guidance. The fine-tuned LLMs are capable of controlling systems by generating sufficiently accurate outputs that are multi-dimensional vectors with up to 10 significant digits. We show that for several problems the amount of data required to perform fine-tuning is smaller than what is generally required of traditional deep neural networks (DNNs), and that fine-tuned LLMs are good at generalizing outside of the training dataset. Further, the same LLM can be fine-tuned with data from different problems, with only minor performance degradation with respect to LLMs trained for a single application. This work is intended as a first step towards the development of a general space systems controller.
Tight Constraint Prediction of Six-Degree-of-Freedom Transformer-based Powered Descent Guidance
Briden, Julia, Gurga, Trey, Johnson, Breanna, Cauligi, Abhishek, Linares, Richard
This work introduces Transformer-based Successive Convexification (T-SCvx), an extension of Transformer-based Powered Descent Guidance (T-PDG), generalizable for efficient six-degree-of-freedom (DoF) fuel-optimal powered descent trajectory generation. Our approach significantly enhances the sample efficiency and solution quality for nonconvex-powered descent guidance by employing a rotation invariant transformation of the sampled dataset. T-PDG was previously applied to the 3-DoF minimum fuel powered descent guidance problem, improving solution times by up to an order of magnitude compared to lossless convexification (LCvx). By learning to predict the set of tight or active constraints at the optimal control problem's solution, Transformer-based Successive Convexification (T-SCvx) creates the minimal reduced-size problem initialized with only the tight constraints, then uses the solution of this reduced problem to warm-start the direct optimization solver. 6-DoF powered descent guidance is known to be challenging to solve quickly and reliably due to the nonlinear and non-convex nature of the problem, the discretization scheme heavily influencing solution validity, and reference trajectory initialization determining algorithm convergence or divergence. Our contributions in this work address these challenges by extending T-PDG to learn the set of tight constraints for the successive convexification (SCvx) formulation of the 6-DoF powered descent guidance problem. In addition to reducing the problem size, feasible and locally optimal reference trajectories are also learned to facilitate convergence from the initial guess. T-SCvx enables onboard computation of real-time guidance trajectories, demonstrated by a 6-DoF Mars powered landing application problem.
Diffusion Policies for Generative Modeling of Spacecraft Trajectories
Briden, Julia, Johnson, Breanna, Linares, Richard, Cauligi, Abhishek
Despite its promise and the tremendous advances in nonlinear optimization solvers in recent years, trajectory optimization has primarily been constrained to offline usage due to the limited compute capabilities of radiation hardened flight computers [3]. However, with a flurry of proposed mission concepts that call for increasingly greater on-board autonomy [4], bridging this gap in the state-of-practice is necessary to allow for scaling current trajectory design techniques for future missions. Recently, researchers have turned to machine learning and data-driven techniques as a promising method for reducing the runtimes necessary for solving challenging constrained optimization problems [5, 6]. Such approaches entail learning what is known as the problem-to-solution mapping between the problem parameters that vary between repeated instances of solving the trajectory optimization problem to the full optimization solution and these works typically use a Deep Neural Network (DNN) to model this mapping [7-9]. Given parameters of new instances of the trajectory optimization problem, this problem-to-solution mapping can be used online to yield candidate trajectories to warm start the nonlinear optimization solver and this warm start can enable significant solution speed ups. One shortcoming of these aforementioned data-driven approaches is that they have limited scope of use and the learned problem-to-solution mapping only applies for one specific trajectory optimization formulation. With a change to the mission design specifications that yields, e.g., a different optimization constraint, a new problem-to-solution mapping has to be learned offline and this necessitates generating a new dataset of solved trajectory optimization problems. To this end, our work explores the use of compositional diffusion modeling to allow for generalizable learning of the problem-to-solution mapping and equip mission designers with the ability to interleave different learned models to satisfy a rich set of trajectory design specifications. Compositional diffusion modeling enables training of a model to both sample and plan from.
Compositional Diffusion Models for Powered Descent Trajectory Generation with Flexible Constraints
Briden, Julia, Du, Yilun, Zucchelli, Enrico M., Linares, Richard
This work introduces TrajDiffuser, a compositional diffusion-based flexible and concurrent trajectory generator for 6 degrees of freedom powered descent guidance. TrajDiffuser is a statistical model that learns the multi-modal distributions of a dataset of simulated optimal trajectories, each subject to only one or few constraints that may vary for different trajectories. During inference, the trajectory is generated simultaneously over time, providing stable long-horizon planning, and constraints can be composed together, increasing the model's generalizability and decreasing the training data required. The generated trajectory is then used to initialize an optimizer, increasing its robustness and speed.
Constraint-Informed Learning for Warm Starting Trajectory Optimization
Briden, Julia, Choi, Changrak, Yun, Kyongsik, Linares, Richard, Cauligi, Abhishek
Future spacecraft and surface robotic missions require increasingly capable autonomy stacks for exploring challenging and unstructured domains and trajectory optimization will be a cornerstone of such autonomy stacks. However, the nonlinear optimization solvers required remain too slow for use on relatively resource constrained flight-grade computers. In this work, we turn towards amortized optimization, a learning-based technique for accelerating optimization run times, and present TOAST: Trajectory Optimization with Merit Function Warm Starts. Offline, using data collected from a simulation, we train a neural network to learn a mapping to the full primal and dual solutions given the problem parameters. Crucially, we build upon recent results from decision-focused learning and present a set of decision-focused loss functions using the notion of merit functions for optimization problems. We show that training networks with such constraint-informed losses can better encode the structure of the trajectory optimization problem and jointly learn to reconstruct the primal-dual solution while also yielding improved constraint satisfaction. Through numerical experiments on a Lunar rover problem, we demonstrate that TOAST outperforms benchmark approaches in terms of both computation times and network prediction constraint satisfaction.
Improving Computational Efficiency for Powered Descent Guidance via Transformer-based Tight Constraint Prediction
Briden, Julia, Gurga, Trey, Johnson, Breanna, Cauligi, Abhishek, Linares, Richard
In this work, we present Transformer-based Powered Descent Guidance (T-PDG), a scalable algorithm for reducing the computational complexity of the direct optimization formulation of the spacecraft powered descent guidance problem. T-PDG uses data from prior runs of trajectory optimization algorithms to train a transformer neural network, which accurately predicts the relationship between problem parameters and the globally optimal solution for the powered descent guidance problem. The solution is encoded as the set of tight constraints corresponding to the constrained minimum-cost trajectory and the optimal final time of landing. By leveraging the attention mechanism of transformer neural networks, large sequences of time series data can be accurately predicted when given only the spacecraft state and landing site parameters. When applied to the real problem of Mars powered descent guidance, T-PDG reduces the time for computing the 3 degree of freedom fuel-optimal trajectory, when compared to lossless convexification, from an order of 1-8 seconds to less than 500 milliseconds. A safe and optimal solution is guaranteed by including a feasibility check in T-PDG before returning the final trajectory.
Transformer-based Atmospheric Density Forecasting
Briden, Julia, Siew, Peng Mun, Rodriguez-Fernandez, Victor, Linares, Richard
As the peak of the solar cycle approaches in 2025 and the ability of a single geomagnetic storm to significantly alter the orbit of Resident Space Objects (RSOs), techniques for atmospheric density forecasting are vital for space situational awareness. While linear data-driven methods, such as dynamic mode decomposition with control (DMDc), have been used previously for forecasting atmospheric density, deep learning-based forecasting has the ability to capture nonlinearities in data. By learning multiple layer weights from historical atmospheric density data, long-term dependencies in the dataset are captured in the mapping between the current atmospheric density state and control input to the atmospheric density state at the next timestep. This work improves upon previous linear propagation methods for atmospheric density forecasting, by developing a nonlinear transformer-based architecture for atmospheric density forecasting. Empirical NRLMSISE-00 and JB2008, as well as physics-based TIEGCM atmospheric density models are compared for forecasting with DMDc and with the transformer-based propagator.