A self-consistent theory of Gaussian Processes captures feature learning effects infinite CNNs
–Neural Information Processing Systems
Despite its theoretical appeal, this viewpoint lacks a crucial ingredient of deep learning in finite DNNs, laying at the heart of their success -- feature learning . Here we consider DNNs trained with noisy gradient descent on a large training set and derive a self-consistent Gaussian Process theory accounting for strong finite-DNN and feature learning effects.
Neural Information Processing Systems
Nov-15-2025, 11:32:32 GMT