Nagarajan, Viswanath
Lower Bound on the Greedy Approximation Ratio for Adaptive Submodular Cover
Harris, Blake, Nagarajan, Viswanath
Adaptive-submodularity is a widely used framework in stochastic optimization and machine learning [GK11, GK17, EKM21, ACN22]. Here, an algorithm makes sequential decisions while (partially) observing uncertainty. We study a basic problem in this context: covering an adaptive-submodular function at the minimum expected cost. We show that the natural greedy algorithm for this problem has approximation ratio at least 1.3 (1 + ln Q), where Q is the maximal function value. This is in contrast to special cases such as deterministic submodular cover or (independent) stochastic submodular cover, where the greedy algorithm achieves a tight (1 + ln Q) approximation ratio.
Semi-Bandit Learning for Monotone Stochastic Optimization
Agarwal, Arpit, Ghuge, Rohan, Nagarajan, Viswanath
Stochastic optimization is a widely used approach for optimization under uncertainty, where uncertain input parameters are modeled by random variables. Exact or approximation algorithms have been obtained for several fundamental problems in this area. However, a significant limitation of this approach is that it requires full knowledge of the underlying probability distributions. Can we still get good (approximation) algorithms if these distributions are unknown, and the algorithm needs to learn them through repeated interactions? In this paper, we resolve this question for a large class of "monotone" stochastic problems, by providing a generic online learning algorithm with $\sqrt{T \log T}$ regret relative to the best approximation algorithm (under known distributions). Importantly, our online algorithm works in a semi-bandit setting, where in each period, the algorithm only observes samples from the r.v.s that were actually probed. Our framework applies to several fundamental problems in stochastic optimization such as prophet inequality, Pandora's box, stochastic knapsack, stochastic matchings and stochastic submodular optimization.
Optimal Decision Tree with Noisy Outcomes
Jia, Su, Navidi, Fatemeh, Nagarajan, Viswanath, Ravi, R.
In pool-based active learning, the learner is given an unlabeled data set and aims to efficiently learn the unknown hypothesis by querying the labels of the data points. This can be formulated as the classical Optimal Decision Tree (ODT) problem: Given a set of tests, a set of hypotheses, and an outcome for each pair of test and hypothesis, our objective is to find a low-cost testing procedure (i.e., decision tree) that identifies the true hypothesis. This optimization problem has been extensively studied under the assumption that each test generates a deterministic outcome. However, in numerous applications, for example, clinical trials, the outcomes may be uncertain, which renders the ideas from the deterministic setting invalid. In this work, we study a fundamental variant of the ODT problem in which some test outcomes are noisy, even in the more general case where the noise is persistent, i.e., repeating a test gives the same noisy output. Our approximation algorithms provide guarantees that are nearly best possible and hold for the general case of a large number of noisy outcomes per test or per hypothesis where the performance degrades continuously with this number. We numerically evaluated our algorithms for identifying toxic chemicals and learning linear classifiers, and observed that our algorithms have costs very close to the information-theoretic minimum.