Penwarden, Michael
A unified scalable framework for causal sweeping strategies for Physics-Informed Neural Networks (PINNs) and their temporal decompositions
Penwarden, Michael, Jagtap, Ameya D., Zhe, Shandian, Karniadakis, George Em, Kirby, Robert M.
Physics-informed neural networks (PINNs) as a means of solving partial differential equations (PDE) have garnered much attention in the Computational Science and Engineering (CS&E) world. However, a recent topic of interest is exploring various training (i.e., optimization) challenges - in particular, arriving at poor local minima in the optimization landscape results in a PINN approximation giving an inferior, and sometimes trivial, solution when solving forward time-dependent PDEs with no data. This problem is also found in, and in some sense more difficult, with domain decomposition strategies such as temporal decomposition using XPINNs. We furnish examples and explanations for different training challenges, their cause, and how they relate to information propagation and temporal decomposition. We then propose a new stacked-decomposition method that bridges the gap between time-marching PINNs and XPINNs. We also introduce significant computational speed-ups by using transfer learning concepts to initialize subnetworks in the domain and loss tolerance-based propagation for the subdomains. Finally, we formulate a new time-sweeping collocation point algorithm inspired by the previous PINNs causality literature, which our framework can still describe, and provides a significant computational speed-up via reduced-cost collocation point segmentation. The proposed methods form our unified framework, which overcomes training challenges in PINNs and XPINNs for time-dependent PDEs by respecting the causality in multiple forms and improving scalability by limiting the computation required per optimization iteration. Finally, we provide numerical results for these methods on baseline PDE problems for which unmodified PINNs and XPINNs struggle to train.
Meta Learning of Interface Conditions for Multi-Domain Physics-Informed Neural Networks
Li, Shibo, Penwarden, Michael, Xu, Yiming, Tillinghast, Conor, Narayan, Akil, Kirby, Robert M., Zhe, Shandian
Physics-informed neural networks (PINNs) are emerging as popular mesh-free solvers for partial differential equations (PDEs). Recent extensions decompose the domain, apply different PINNs to solve the problem in each subdomain, and stitch the subdomains at the interface. Thereby, they can further alleviate the problem complexity, reduce the computational cost, and allow parallelization. However, the performance of multi-domain PINNs is sensitive to the choice of the interface conditions. While quite a few conditions have been proposed, there is no suggestion about how to select the conditions according to specific problems. To address this gap, we propose META Learning of Interface Conditions (METALIC), a simple, efficient yet powerful approach to dynamically determine appropriate interface conditions for solving a family of parametric PDEs. Specifically, we develop two contextual multi-arm bandit (MAB) models. The first one applies to the entire training course, and online updates a Gaussian process (GP) reward that given the PDE parameters and interface conditions predicts the performance. We prove a sub-linear regret bound for both UCB and Thompson sampling, which in theory guarantees the effectiveness of our MAB. The second one partitions the training into two stages, one is the stochastic phase and the other deterministic phase; we update a GP reward for each phase to enable different condition selections at the two stages to further bolster the flexibility and performance. We have shown the advantage of METALIC on four bench-mark PDE families.
Neural Operator Learning for Ultrasound Tomography Inversion
Dai, Haocheng, Penwarden, Michael, Kirby, Robert M., Joshi, Sarang
Neural operator learning as a means of mapping between complex function spaces has garnered significant attention in the field of computational science and engineering (CS&E). In this paper, we apply Neural operator learning to the time-of-flight ultrasound computed tomography (USCT) problem. We learn the mapping between time-of-flight (TOF) data and the heterogeneous sound speed field using a full-wave solver to generate the training data. This novel application of operator learning circumnavigates the need to solve the computationally intensive iterative inverse problem. The operator learns the non-linear mapping offline and predicts the heterogeneous sound field with a single forward pass through the model. This is the first time operator learning has been used for ultrasound tomography and is the first step in potential real-time predictions of soft tissue distribution for tumor identification in beast imaging.
Deep neural operators can serve as accurate surrogates for shape optimization: A case study for airfoils
Shukla, Khemraj, Oommen, Vivek, Peyvan, Ahmad, Penwarden, Michael, Bravo, Luis, Ghoshal, Anindya, Kirby, Robert M., Karniadakis, George Em
Neural networks that solve regression problems map input data to output data, whereas neural operators map functions to functions. This recent paradigm shift in perspective, starting with the original paper on the deep operator network or DeepONet [1, 2], provides a new modeling capability that is very useful in engineering - that is, the ability to replace very complex and computational resource-taxing multiphysics systems with neural operators that can provide functional outputs in real-time. Specifically, unlike other physics-informed neural networks (PINNs) [3] that require optimization during training and testing, a DeepONet does not require any optimization during inference, hence it can be used in realtime forecasting, including design, autonomy, control, etc. An architectural diagram of a DeepONet with the commonly used nomenclature for its components is shown in Figure 1. DeepONets can take a multi-fidelity or multi-modal input [4, 5, 6, 7, 8] in the branch network and can use an independent network as the trunk, a network that represents the output space, e.g. in space-time coordinates or in parametric space in a continuous fashion. In some sense, DeepONets can be used as surrogates in a similar fashion as reduced order models (ROMs) [9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19]. However, unlike ROMs, they are over-parametrized which leads to both generalizability and robustness to noise that is not possible with ROMs, see the recent work of [20].
A Metalearning Approach for Physics-Informed Neural Networks (PINNs): Application to Parameterized PDEs
Penwarden, Michael, Zhe, Shandian, Narayan, Akil, Kirby, Robert M.
Physics-informed neural networks (PINNs) as a means of discretizing partial differential equations (PDEs) are garnering much attention in the Computational Science and Engineering (CS&E) world. At least two challenges exist for PINNs at present: an understanding of accuracy and convergence characteristics with respect to tunable parameters and identification of optimization strategies that make PINNs as efficient as other computational science tools. The cost of PINNs training remains a major challenge of Physics-informed Machine Learning (PiML) - and, in fact, machine learning (ML) in general. This paper is meant to move towards addressing the latter through the study of PINNs on new tasks, for which parameterized PDEs provides a good testbed application as tasks can be easily defined in this context. Following the ML world, we introduce metalearning of PINNs with application to parameterized PDEs. By introducing metalearning and transfer learning concepts, we can greatly accelerate the PINNs optimization process. We present a survey of model-agnostic metalearning, and then discuss our model-aware metalearning applied to PINNs as well as implementation considerations and algorithmic complexity. We then test our approach on various canonical forward parameterized PDEs that have been presented in the emerging PINNs literature.
Multifidelity Modeling for Physics-Informed Neural Networks (PINNs)
Penwarden, Michael, Zhe, Shandian, Narayan, Akil, Kirby, Robert M.
Multifidelity simulation methodologies are often used in an attempt to judiciously combine low-fidelity and high-fidelity simulation results in an accuracy-increasing, cost-saving way. Candidates for this approach are simulation methodologies for which there are fidelity differences connected with significant computational cost differences. Physics-informed Neural Networks (PINNs) are candidates for these types of approaches due to the significant difference in training times required when different fidelities (expressed in terms of architecture width and depth as well as optimization criteria) are employed. In this paper, we propose a particular multifidelity approach applied to PINNs that exploits low-rank structure. We demonstrate that width, depth, and optimization criteria can be used as parameters related to model fidelity, and show numerical justification of cost differences in training due to fidelity parameter choices. We test our multifidelity scheme on various canonical forward PDE models that have been presented in the emerging PINNs literature.