Exact natural gradient in deep linear networks and its application to the nonlinear case
Alberto Bernacchia, Mate Lengyel, Guillaume Hennequin
–Neural Information Processing Systems
Stochastic gradient descent (SGD) is used ubiquitously to train deep neural networks, due to its low computational cost and ease of implementation.
Neural Information Processing Systems
Nov-18-2025, 02:08:16 GMT
- Country:
- Europe
- Hungary > Budapest
- Budapest (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.14)
- Hungary > Budapest
- North America > Canada
- Europe
- Technology: