Goto

Collaborating Authors

 assumption2


A Direct Approach for Handling Contextual Bandits with Latent State Dynamics

Li, Zhen, Stoltz, Gilles

arXiv.org Machine Learning

We revisit the finite-armed linear bandit model by Nelson et al. (2022), where contexts and rewards are governed by a finite hidden Markov chain. Nelson et al. (2022) approach this model by a reduction to linear contextual bandits; but to do so, they actually introduce a simplification in which rewards are linear functions of the posterior probabilities over the hidden states given the observed contexts, rather than functions of the hidden states themselves. Their analysis (but not their algorithm) also does not take into account the estimation of the HMM parameters, and only tackles expected, not high-probability, bounds, which suffer in addition from unnecessary complex dependencies on the model (like reward gaps). We instead study the more natural model incorporating direct dependencies in the hidden states (on top of dependencies on the observed contexts, as is natural for contextual bandits) and also obtain stronger, high-probability, regret bounds for a fully adaptive strategy that estimates HMM parameters online. These bounds do not depend on the reward functions and only depend on the model through the estimation of the HMM parameters.



OntheEffectiveNumberofLinearRegionsinShallow UnivariateReLUNetworks: ConvergenceGuarantees andImplicitBias

Neural Information Processing Systems

Howeverwhat is perhaps more surprising, is that in stark contrast to our classic understanding of generalization in machine learning models, this does not seem to degrade the generalization capabilities of the learned model in spite of the significant increase in its capacity.



d3e6cd9f66f2c1d3840ade4161cf7406-Paper.pdf

Neural Information Processing Systems

Our bounds hold ininfinite-dimensional spaces, thereby showing that finer and finer discretizations do not make this learning problemharder.




OnReward-FreeReinforcementLearningwith LinearFunctionApproximation

Neural Information Processing Systems

During the exploration phase, an agent collects samples without using a pre-specified reward function. After the exploration phase, a reward function is given, and the agent uses samples collected during the exploration phase to computeanear-optimalpolicy.


EscapingSaddle-PointFasterunder Interpolation-likeConditions

Neural Information Processing Systems

One of the fundamental aspects of over-parametrized models is that they are capable of interpolating the training data. We show that, under interpolation-like assumptions satisfied by the stochastic gradients in an overparametrization setting, thefirst-order oracle complexityofPerturbed Stochastic Gradient Descent (PSGD) algorithm toreach an -local-minimizer,matches the corresponding deterministic rateof O(1/2).