On the Convergence of Loss and Uncertainty-based Active Learning Algorithms
–Neural Information Processing Systems
We investigate the convergence rates and data sample sizes required for training a machine learning model using a stochastic gradient descent (SGD) algorithm, where data points are sampled based on either their loss value or uncertainty value. These training methods are particularly relevant for active learning and data subset selection problems. For SGD with a constant step size update, we present convergence results for linear classifiers and linearly separable datasets using squared hinge loss and similar training loss functions.
Neural Information Processing Systems
Feb-18-2026, 09:39:59 GMT
- Country:
- Africa > Ethiopia
- Addis Ababa > Addis Ababa (0.04)
- Europe > Slovenia
- Drava > Municipality of Benedikt > Benedikt (0.04)
- North America > United States
- California > San Francisco County
- San Francisco (0.14)
- Georgia > Fulton County
- Atlanta (0.04)
- Louisiana > Orleans Parish
- New Orleans (0.04)
- Wisconsin > Dane County
- Madison (0.04)
- California > San Francisco County
- Africa > Ethiopia
- Genre:
- Research Report
- Experimental Study (1.00)
- New Finding (1.00)
- Research Report
- Technology: