PDE-Transformer: Efficient and Versatile Transformers for Physics Simulations
Holzschuh, Benjamin, Liu, Qiang, Kohl, Georg, Thuerey, Nils
–arXiv.org Artificial Intelligence
We introduce PDE-Transformer, an improved transformer-based architecture for surrogate modeling of physics simulations on regular grids. We combine recent architectural improvements of diffusion transformers with adjustments specific for large-scale simulations to yield a more scalable and versatile general-purpose transformer architecture, which can be used as the backbone for building large-scale foundation models in physical sciences. We demonstrate that our proposed architecture outperforms state-of-the-art transformer architectures for computer vision on a large dataset of 16 different types of PDEs. We propose to embed different physical channels individually as spatio-temporal tokens, which interact via channel-wise self-attention. This helps to maintain a consistent information density of tokens when learning multiple types of PDEs simultaneously. We demonstrate that our pre-trained models achieve improved performance on several challenging downstream tasks compared to training from scratch and also beat other foundation model architectures for physics simulations.
arXiv.org Artificial Intelligence
Jun-2-2025
- Country:
- Africa > Rwanda
- Europe
- North America
- Canada > British Columbia
- Vancouver (0.04)
- Costa Rica > San José Province
- San José (0.04)
- United States
- Arizona > Pima County
- Tucson (0.04)
- Hawaii > Honolulu County
- Honolulu (0.04)
- Louisiana > Orleans Parish
- New Orleans (0.04)
- Minnesota > Hennepin County
- Minneapolis (0.14)
- Arizona > Pima County
- Canada > British Columbia
- Genre:
- Research Report (0.64)
- Industry:
- Health & Medicine (0.46)
- Technology: