Scaled Least Squares Estimator for GLMs in Large-Scale Problems
Erdogdu, Murat A., Dicker, Lee H., Bayati, Mohsen
–Neural Information Processing Systems
We study the problem of efficiently estimating the coefficients of generalized linear models (GLMs) in the large-scale setting where the number of observations n is much larger than the number of predictors p, i.e. n p 1. We show that in GLMs with random (not necessarily Gaussian) design, the GLM coefficients are approximately proportional to the corresponding ordinary least squares (OLS) coefficients. Usingthis relation, we design an algorithm that achieves the same accuracy as the maximum likelihood estimator (MLE) through iterations that attain up to a cubic convergence rate, and that are cheaper than any batch optimization algorithm by at least a factor of O(p). We provide theoretical guarantees for our algorithm, and analyze the convergence behavior in terms of data dimensions. Finally, we demonstrate the performance of our algorithm through extensive numerical studies on large-scale real and synthetic datasets, and show that it achieves the highest performance compared to several other widely used optimization algorithms.
Neural Information Processing Systems
Dec-31-2016
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe
- Spain > Catalonia
- Barcelona Province > Barcelona (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Oxfordshire > Oxford (0.04)
- Spain > Catalonia
- North America > United States
- California > Santa Clara County
- Palo Alto (0.04)
- Indiana > Hamilton County
- Fishers (0.04)
- California > Santa Clara County
- Asia > Middle East
- Genre:
- Research Report (0.49)
- Technology: