Neural Conservation Laws: A Divergence-Free Perspective
Richter-Powell, Jack, Lipman, Yaron, Chen, Ricky T. Q.
–arXiv.org Artificial Intelligence
We investigate the parameterization of deep neural networks that by design satisfy the continuity equation, a fundamental conservation law. This is enabled by the observation that any solution of the continuity equation can be represented as a divergence-free vector field. We hence propose building divergence-free neural networks through the concept of differential forms, and with the aid of automatic differentiation, realize two practical constructions. As a result, we can parameterize pairs of densities and vector fields that always exactly satisfy the continuity equation, foregoing the need for extra penalty methods or expensive numerical simulation. Furthermore, we prove these models are universal and so can be used to represent any divergence-free vector field.
arXiv.org Artificial Intelligence
Dec-11-2022