Goto

Collaborating Authors

 margin




Learning Noisy Halfspaces with a Margin: Massart is No Harder than Random

Neural Information Processing Systems

We study the problem of PAC learning \gamma -margin halfspaces with Massart noise. We propose a simple proper learning algorithm, the Perspectron, that has sample complexity \widetilde{O}((\epsilon\gamma) {-2}) and achieves classification error at most \eta \epsilon where \eta is the Massart noise rate. Prior works (DGT19, CKMY20) came with worse sample complexity guarantees (in both \epsilon and \gamma) or could only handle random classification noise (DDKWZ23,KITBMV23)--- a much milder noise assumption. We also show that our results extend to the more challenging setting of learning generalized linear models with a known link function under Massart noise, achieving a similar sample complexity to the halfspace case. This significantly improves upon the prior state-of-the-art in this setting due to CKMY20, who introduced this model.


Online Bidding Algorithms with Strict Return on Spend (ROS) Constraint

Vaze, Rahul, Sinha, Abhishek

arXiv.org Artificial Intelligence

Auto-bidding problem under a strict return-on-spend constraint (ROSC) is considered, where an algorithm has to make decisions about how much to bid for an ad slot depending on the revealed value, and the hidden allocation and payment function that describes the probability of winning the ad-slot depending on its bid. The objective of an algorithm is to maximize the expected utility (product of ad value and probability of winning the ad slot) summed across all time slots subject to the total expected payment being less than the total expected utility, called the ROSC. A (surprising) impossibility result is derived that shows that no online algorithm can achieve a sub-linear regret even when the value, allocation and payment function are drawn i.i.d. from an unknown distribution. The problem is non-trivial even when the revealed value remains constant across time slots, and an algorithm with regret guarantee that is optimal up to logarithmic factor is derived.


Book Reviews

AI Magazine

Parametric tests are only valid if the data satisfy certain assumptions. If these assumptions hold, they will, however, typically give more accurate results. The analysis of statistical learning theory has very much the flavor of a nonparametric statistical test. The weakness of pac, therefore, is that its results must hold true even in worst-case distributions. There is, however, a new twist to this story in that the more recent pacstyle results are able to take account of observed attributes of the function that has been chosen by the learner, for example, its margin on the training set.


Support Vector Machines and Kernel Methods

AI Magazine

Kernel methods, a new generation of learning algorithms, utilize techniques from optimization, statistics, and functional analysis to achieve maximal generality, flexibility, and performance. These algorithms are different from earlier techniques used in machine learning in many respects: For example, they are explicitly based on a theoretical model of learning rather than on loose analogies with natural learning systems or other heuristics. They come with theoretical guarantees about their performance and have a modular design that makes it possible to separately implement and analyze their components. They are not affected by the problem of local minima because their training amounts to convex optimization. In the last decade, a sizable community of theoreticians and practitioners has formed around these methods, and a number of practical applications have been realized.