Faster Boosting with Smaller Memory
Alafate, Julaiti, Freund, Yoav
The two state-of-the-art implementations of boosted trees: XGBoost and LightGBM, can process large training sets extremely fast. However, this performance requires that memory size is sufficient to hold a 2-3 multiple of the training set size. This paper presents an alternative approach to implementing boosted trees. which achieves a significant speedup over XGBoost and LightGBM, especially when memory size is small. This is achieved using a combination of two techniques: early stopping and stratified sampling, which are explained and analyzed in the paper. We describe our implementation and present experimental results to support our claims.
Jan-25-2019
- Country:
- Asia > Japan (0.14)
- North America > United States
- California > San Diego County
- San Diego (0.04)
- Massachusetts > Middlesex County
- Cambridge (0.04)
- New York > New York County
- New York City (0.04)
- California > San Diego County
- Oceania > New Zealand
- North Island > Waikato (0.04)
- Genre:
- Research Report (0.50)
- Technology: