Debiasing Conditional Stochastic Optimization
He, Lie, Kasiviswanathan, Shiva Prasad
In this paper, we study the conditional stochastic optimization (CSO) problem which covers a variety of applications including portfolio selection, reinforcement learning, robust learning, causal inference, etc. The sample-averaged gradient of the CSO objective is biased due to its nested structure, and therefore requires a high sample complexity for convergence. We introduce a general stochastic extrapolation technique that effectively reduces the bias. We show that for nonconvex smooth objectives, combining this extrapolation with variance reduction techniques can achieve a significantly better sample complexity than the existing bounds. Additionally, we develop new algorithms for the finite-sum variant of the CSO problem that also significantly improve upon existing results. Finally, we believe that our debiasing technique has the potential to be a useful tool for addressing similar challenges in other stochastic optimization problems.
Dec-3-2023
- Country:
- Europe (0.92)
- North America
- Canada > Quebec (0.14)
- United States > New York (0.14)
- Genre:
- Research Report > New Finding (0.67)
- Technology: