DS-MLR: Exploiting Double Separability for Scaling up Distributed Multinomial Logistic Regression
Raman, Parameswaran, Srinivasan, Sriram, Matsushima, Shin, Zhang, Xinhua, Yun, Hyokun, Vishwanathan, S. V. N.
Scaling multinomial logistic regression to datasets with very large number of data points and classes has not been trivial. This is primarily because one needs to compute the log-partition function on every data point. This makes distributing the computation hard. In this paper, we present a distributed stochastic gradient descent based optimization method (DS-MLR) for scaling up multinomial logistic regression problems to massive scale datasets without hitting any storage constraints on the data and model parameters. Our algorithm exploits double-separability, an attractive property we observe in the objective functions of several models in machine learning, that allows us to achieve both data as well as model parallelism simultaneously. In addition to being parallelizable, our algorithm can also easily be made non-blocking and asynchronous. We demonstrate the effectiveness of DS-MLR empirically on several real-world datasets, the largest being a reddit dataset created out of 1.7 billion user comments, where the data and parameter sizes are 228 GB and 358 GB respectively.
Feb-14-2018
- Country:
- Europe > United Kingdom
- England > Cambridgeshire > Cambridge (0.14)
- North America > United States (0.28)
- Europe > United Kingdom
- Genre:
- Research Report
- Experimental Study (0.82)
- New Finding (0.82)
- Research Report
- Technology: