Correcting auto-differentiation in neural-ODE training
Xu, Yewei, Chen, Shi, Li, Qin, Wright, Stephen J.
–arXiv.org Artificial Intelligence
Does the use of auto-differentiation yield reasonable updates to deep neural networks that represent neural ODEs? Through mathematical analysis and numerical evidence, we find that when the neural network employs high-order forms to approximate the underlying ODE flows (such as the Linear Multistep Method (LMM)), brute-force computation using auto-differentiation often produces non-converging artificial oscillations. In the case of Leapfrog, we propose a straightforward post-processing technique that effectively eliminates these oscillations, rectifies the gradient computation and thus respects the updates of the underlying flow.
arXiv.org Artificial Intelligence
Jun-3-2023
- Country:
- North America > United States > Wisconsin (0.29)
- Genre:
- Research Report (0.50)
- Industry:
- Energy (0.68)
- Technology: