Accelerated Training of Physics-Informed Neural Networks (PINNs) using Meshless Discretizations

Neural Information Processing Systems 

Additionally, though traditional PINNs (vanilla-PINNs) are typically stored and trained in 32-bit floating-point (fp32) on the GPU, we show that for DT -PINNs, using fp64 on the GPU leads to significantly faster training times than fp32 vanilla-PINNs with comparable accuracy.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found