Reviews: Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent

Neural Information Processing Systems 

The paper was proofread, well-structured, and very clear. The experiments were clearly described in detail, and provided relevant results. Below we outline some detailed comments of the results. In particular, Chizat and Bach prove that the training of an NTK parameterized network is closely modeled by "lazy training" (their terminology for a linearized model). This paper is not referenced in the related work section.