Newton-ADMM: A Distributed GPU-Accelerated Optimizer for Multiclass Classification Problems
Fang, Chih-Hao, Kylasa, Sudhir B, Roosta, Fred, Mahoney, Michael W., Grama, Ananth
First-order optimization methods, such as stochastic gradient descent (SGD) and its variants, are widely used in machine learning applications due to their simplicity and low per-iteration costs. However, they often require larger numbers of iterations, with associated communication costs in distributed environments. In contrast, Newton-type methods, while having higher per-iteration costs, typically require a significantly smaller number of iterations, which directly translates to reduced communication costs. In this paper, we present a novel distributed optimizer for classification problems, which integrates a GPU-accelerated Newton-type solver with the global consensus formulation of Alternating Direction of Method Multipliers (ADMM). By leveraging the communication efficiency of ADMM, GPU-accelerated inexact-Newton solver, and an effective spectral penalty parameter selection strategy, we show that our proposed method (i) yields better generalization performance on several classification problems; (ii) significantly outperforms state-of-the-art methods in distributed time to solution; and (iii) offers better scaling on large distributed platforms.
Feb-4-2020
- Country:
- North America > United States
- California > Alameda County
- Berkeley (0.04)
- Indiana > Tippecanoe County
- Lafayette (0.04)
- California > Alameda County
- Oceania > Australia
- Queensland (0.04)
- North America > United States
- Genre:
- Research Report (1.00)
- Technology: