Safe Adaptive Importance Sampling
Stich, Sebastian U., Raj, Anant, Jaggi, Martin
–Neural Information Processing Systems
Importance sampling has become an indispensable strategy to speed up optimization algorithms for large-scale applications. Improved adaptive variants -- using importance values defined by the complete gradient information which changes during optimization -- enjoy favorable theoretical properties, but are typically computationally infeasible. In this paper we propose an efficient approximation of gradient-based sampling, which is based on safe bounds on the gradient. The proposed sampling distribution is (i) provably the \emph{best sampling} with respect to the given bounds, (ii) always better than uniform sampling and fixed importance sampling and (iii) can efficiently be computed -- in many applications at negligible extra cost. The proposed sampling scheme is generic and can easily be integrated into existing algorithms. In particular, we show that coordinate-descent (CD) and stochastic gradient descent (SGD) can enjoy significant a speed-up under the novel scheme. The proven efficiency of the proposed sampling is verified by extensive numerical testing.
Neural Information Processing Systems
Dec-31-2017
- Country:
- Europe
- Germany > Baden-Württemberg
- Tübingen Region > Tübingen (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Germany > Baden-Württemberg
- North America > United States
- California > Los Angeles County
- Long Beach (0.04)
- New York (0.04)
- California > Los Angeles County
- Europe
- Genre:
- Research Report > New Finding (0.68)
- Technology: