Neural Message Passing Induced by Energy-Constrained Diffusion
Wu, Qitian, Wipf, David, Yan, Junchi
–arXiv.org Artificial Intelligence
Learning representations for structured data with certain geometries (observed or unobserved) is a fundamental challenge, wherein message passing neural networks (MPNNs) have become a de facto class of model solutions. In this paper, we propose an energy-constrained diffusion model as a principled interpretable framework for understanding the mechanism of MPNNs and navigating novel architectural designs. The model, inspired by physical systems, combines the inductive bias of diffusion on manifolds with layer-wise constraints of energy minimization. As shown by our analysis, the diffusion operators have a one-to-one correspondence with the energy functions implicitly descended by the diffusion process, and the finite-difference iteration for solving the energy-constrained diffusion system induces the propagation layers of various types of MPNNs operated on observed or latent structures. On top of these findings, we devise a new class of neural message passing models, dubbed as diffusion-inspired Transformers, whose global attention layers are induced by the principled energy-constrained diffusion. Across diverse datasets ranging from real-world networks to images and physical particles, we show that the new model can yield promising performance for cases where the data structures are observed (as a graph), partially observed or completely unobserved.
arXiv.org Artificial Intelligence
Sep-13-2024
- Country:
- Asia > China
- Europe
- Germany > Baden-Württemberg
- Stuttgart Region > Stuttgart (0.04)
- Hungary (0.04)
- Netherlands (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Germany > Baden-Württemberg
- North America > United States (0.14)
- Genre:
- Research Report > New Finding (0.46)
- Industry:
- Energy (0.47)
- Health & Medicine (0.46)
- Information Technology (0.46)
- Technology: