Why Big Is Not Always Better In Machine Learning
Neural networks are trained to exactly fit the data. Such models usually would be considered as over-fitting, and yet they have managed to obtain high accuracy on test data. It is counter-intuitive -- but it works. This has raised many eyebrows, especially regarding the mathematical foundations of machine learning and their relevance to practitioners. In order to address these contradictions, researchers at OpenAI, in their recent work, double down on this widely believed grand illusion of bigger is better. In this paper, an attempt has been made to reconcile classical understanding and modern practice within a unified performance curve.
Dec-11-2019, 15:15:38 GMT