Goto

Collaborating Authors

 Xiao, Xian


Scalable Back-Propagation-Free Training of Optical Physics-Informed Neural Networks

arXiv.org Artificial Intelligence

Physics-informed neural networks (PINNs) have shown promise in solving partial differential equations (PDEs), with growing interest in their energy-efficient, real-time training on edge devices. Photonic computing offers a potential solution to achieve this goal because of its ultra-high operation speed. However, the lack of photonic memory and the large device sizes prevent training real-size PINNs on photonic chips. This paper proposes a completely back-propagation-free (BP-free) and highly salable framework for training real-size PINNs on silicon photonic platforms. Our approach involves three key innovations: (1) a sparse-grid Stein derivative estimator to avoid the BP in the loss evaluation of a PINN, (2) a dimension-reduced zeroth-order optimization via tensor-train decomposition to achieve better scalability and convergence in BP-free training, and (3) a scalable on-chip photonic PINN training accelerator design using photonic tensor cores. We validate our numerical methods on both low- and high-dimensional PDE benchmarks. Through circuit simulation based on real device parameters, we further demonstrate the significant performance benefit (e.g., real-time training, huge chip area reduction) of our photonic accelerator.


Experimental Demonstration of an Optical Neural PDE Solver via On-Chip PINN Training

arXiv.org Artificial Intelligence

Examples include electromagnetic modeling and thermal analysis of IC chips [1], medical imaging [2], safety verification of autonomous systems [3]. Discretization-based solvers (e.g., finite-difference and finite-element methods) convert a PDE to a large-scale algebraic equation via spatial and temporal discretization. Solving the resulting algebraic equation often requires massive digital resources and run times. Physics-informed neural network (PINN) is a promising discretization-free and unsupervised approach to solve PDEs [4]. PINN uses the residuals of a PDE operator and the boundary/initial conditions to set up a loss function, then minimizes the loss to train a neural network as a global approximation of the PDE solution. However, current PINN training typically needs several to dozens of hours on a powerful GPU, hindering the deployment of an real-time neural PDE solver on edge devices.


Real-Time FJ/MAC PDE Solvers via Tensorized, Back-Propagation-Free Optical PINN Training

arXiv.org Artificial Intelligence

Solving partial differential equations (PDEs) numerically often requires huge computing time, energy cost, and hardware resources in practical applications. This has limited their applications in many scenarios (e.g., autonomous systems, supersonic flows) that have a limited energy budget and require near real-time response. Leveraging optical computing, this paper develops an on-chip training framework for physics-informed neural networks (PINNs), aiming to solve high-dimensional PDEs with fJ/MAC photonic power consumption and ultra-low latency. Despite the ultra-high speed of optical neural networks, training a PINN on an optical chip is hard due to (1) the large size of photonic devices, and (2) the lack of scalable optical memory devices to store the intermediate results of back-propagation (BP). To enable realistic optical PINN training, this paper presents a scalable method to avoid the BP process. We also employ a tensor-compressed approach to improve the convergence and scalability of our optical PINN training. This training framework is designed with tensorized optical neural networks (TONN) for scalable inference acceleration and MZI phase-domain tuning for \textit{in-situ} optimization. Our simulation results of a 20-dim HJB PDE show that our photonic accelerator can reduce the number of MZIs by a factor of $1.17\times 10^3$, with only $1.36$ J and $1.15$ s to solve this equation. This is the first real-size optical PINN training framework that can be applied to solve high-dimensional PDEs.