Exact natural gradient in deep linear networks and its application to the nonlinear case

Alberto Bernacchia, Mate Lengyel, Guillaume Hennequin

Neural Information Processing Systems 

Stochastic gradient descent (SGD) is used ubiquitously to train deep neural networks, due to its low computational cost and ease of implementation.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found