A self-consistent theory of Gaussian Processes captures feature learning effects infinite CNNs

Neural Information Processing Systems 

Despite its theoretical appeal, this viewpoint lacks a crucial ingredient of deep learning in finite DNNs, laying at the heart of their success -- feature learning . Here we consider DNNs trained with noisy gradient descent on a large training set and derive a self-consistent Gaussian Process theory accounting for strong finite-DNN and feature learning effects.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found