Goto

Collaborating Authors

 batchbald



BatchBALD: Efficient and Diverse Batch Acquisition for Deep Bayesian Active Learning

Neural Information Processing Systems

We develop BatchBALD, a tractable approximation to the mutual information between a batch of points and model parameters, which we use as an acquisition function to select multiple informative points jointly for the task of deep Bayesian active learning. BatchBALD is a greedy linear-time $1 - \nicefrac{1}{e}$-approximate algorithm amenable to dynamic programming and efficient caching. We compare BatchBALD to the commonly used approach for batch data acquisition and find that the current approach acquires similar and redundant points, sometimes performing worse than randomly acquiring data. We finish by showing that, using BatchBALD to consider dependencies within an acquisition batch, we achieve new state of the art performance on standard benchmarks, providing substantial data efficiency improvements in batch acquisition.


We thank the reviewers for taking the time to write these thorough reviews and their appreciation of BatchBALD as a

Neural Information Processing Systems

We address reviewer 1, 2 and 3 as R1, R2, R3. R1-(5): We use 25%, 75% quartiles for the shaded areas, see line 147 in the paper. R2 - Originality: Thank you for pointing us to additional relevant related work: we have added citations. We provide additional results on CINIC-10 (top figure, left). We use 50 MC dropout samples, acquisition size 10 and 6 trials.




Reviews: BatchBALD: Efficient and Diverse Batch Acquisition for Deep Bayesian Active Learning

Neural Information Processing Systems

My score remains the same. The methods proposed in the paper elegantly deals with the problem of redundant acquisition when using BALD in a greedy manner. I have a few questions and hope the authors can address them: (1) Does this problem of redundant acquisition only happen when one uses BALD as the score? Intuitively I would think no, as if one uses any score function greedily, regardless of the contribution of the other samples selected in the same batch, one can still end up with a biased batch that can potentially harm training. If this is the case, then why are var-ratios and mean-std outperforming random?


Reviews: BatchBALD: Efficient and Diverse Batch Acquisition for Deep Bayesian Active Learning

Neural Information Processing Systems

The paper proposes BatchBALD, a batch acquisition function for sample selection in active learning. A greedy optimization algorithm is presented for efficient sample selection and BatchBALD score maximization. The reviewers and AC agree that this is an interesting work and that the approach is clearly presented and convincing. In addition the author response satisfactorily addresses the points raised in the reviews.


Big Batch Bayesian Active Learning by Considering Predictive Probabilities

Ober, Sebastian W., Power, Samuel, Diethe, Tom, Moss, Henry B.

arXiv.org Machine Learning

We observe that BatchBALD, a popular acquisition function for batch Bayesian active learning for classification, can conflate epistemic and aleatoric uncertainty, leading to suboptimal performance. Motivated by this observation, we propose to focus on the predictive probabilities, which only exhibit epistemic uncertainty. The result is an acquisition function that not only performs better, but is also faster to evaluate, allowing for larger batches than before.


BatchBALD: Efficient and Diverse Batch Acquisition for Deep Bayesian Active Learning

Neural Information Processing Systems

We develop BatchBALD, a tractable approximation to the mutual information between a batch of points and model parameters, which we use as an acquisition function to select multiple informative points jointly for the task of deep Bayesian active learning. BatchBALD is a greedy linear-time 1 - icefrac{1}{e} -approximate algorithm amenable to dynamic programming and efficient caching. We compare BatchBALD to the commonly used approach for batch data acquisition and find that the current approach acquires similar and redundant points, sometimes performing worse than randomly acquiring data. We finish by showing that, using BatchBALD to consider dependencies within an acquisition batch, we achieve new state of the art performance on standard benchmarks, providing substantial data efficiency improvements in batch acquisition.


Scalable Batch Acquisition for Deep Bayesian Active Learning

Rubashevskii, Aleksandr, Kotova, Daria, Panov, Maxim

arXiv.org Artificial Intelligence

In deep active learning, it is especially important to choose multiple examples to markup at each step to work efficiently, especially on large datasets. At the same time, existing solutions to this problem in the Bayesian setup, such as BatchBALD, have significant limitations in selecting a large number of examples, associated with the exponential complexity of computing mutual information for joint random variables. We, therefore, present the Large BatchBALD algorithm, which gives a well-grounded approximation to the BatchBALD method that aims to achieve comparable quality while being more computationally efficient. We provide a complexity analysis of the algorithm, showing a reduction in computation time, especially for large batches. Furthermore, we present an extensive set of experimental results on image and text data, both on toy datasets and larger ones such as CIFAR-100.