Goto

Collaborating Authors

 assumption2



OntheEffectiveNumberofLinearRegionsinShallow UnivariateReLUNetworks: ConvergenceGuarantees andImplicitBias

Neural Information Processing Systems

Howeverwhat is perhaps more surprising, is that in stark contrast to our classic understanding of generalization in machine learning models, this does not seem to degrade the generalization capabilities of the learned model in spite of the significant increase in its capacity.



d3e6cd9f66f2c1d3840ade4161cf7406-Paper.pdf

Neural Information Processing Systems

Our bounds hold ininfinite-dimensional spaces, thereby showing that finer and finer discretizations do not make this learning problemharder.




OnReward-FreeReinforcementLearningwith LinearFunctionApproximation

Neural Information Processing Systems

During the exploration phase, an agent collects samples without using a pre-specified reward function. After the exploration phase, a reward function is given, and the agent uses samples collected during the exploration phase to computeanear-optimalpolicy.


EscapingSaddle-PointFasterunder Interpolation-likeConditions

Neural Information Processing Systems

One of the fundamental aspects of over-parametrized models is that they are capable of interpolating the training data. We show that, under interpolation-like assumptions satisfied by the stochastic gradients in an overparametrization setting, thefirst-order oracle complexityofPerturbed Stochastic Gradient Descent (PSGD) algorithm toreach an -local-minimizer,matches the corresponding deterministic rateof O(1/2).