Reviews: Wide Feedforward or Recurrent Neural Networks of Any Architecture are Gaussian Processes
–Neural Information Processing Systems
The paper presents a method for collapsing a wide range of operations (convolution, pooling, batchnorm, attention, gating, as well as the inner products for the actual GP Kernel computation) into the matrix multiplication / nonlinearity / linear combination framework; and also a mean field theory of tied weights, which allows a rigorous extension to RNNs as well as a rigorous integration of the forward and backward pass. The results are novel and interesting. This paper had strong overlap with another paper (that was clearly identified by the authors in both submissions), and so the discussion of the tw o papers took place together.
Neural Information Processing Systems
Jan-24-2025, 02:51:42 GMT
- Technology: