corruption distribution
It is not the corruption distribution itself that 2 ultimately generates new, realistic objects; rather, it is the repeated application of the corruption and
We thank the reviewers for the valuable feedback and address specific comments below. We plan to expand Section 2.1 with additional explanation to make the paper more self-contained. Although the samples are not i.i.d., no burn-in or thinning is used. Defining such moves does require some domain expertise. We plan to include updated Guacamol results in the paper.
CRIMED: Lower and Upper Bounds on Regret for Bandits with Unbounded Stochastic Corruption
Agrawal, Shubhada, Mathieu, Timothée, Basu, Debabrota, Maillard, Odalric-Ambrym
We investigate the regret-minimisation problem in a multi-armed bandit setting with arbitrary corruptions. Similar to the classical setup, the agent receives rewards generated independently from the distribution of the arm chosen at each time. However, these rewards are not directly observed. Instead, with a fixed $\varepsilon\in (0,\frac{1}{2})$, the agent observes a sample from the chosen arm's distribution with probability $1-\varepsilon$, or from an arbitrary corruption distribution with probability $\varepsilon$. Importantly, we impose no assumptions on these corruption distributions, which can be unbounded. In this setting, accommodating potentially unbounded corruptions, we establish a problem-dependent lower bound on regret for a given family of arm distributions. We introduce CRIMED, an asymptotically-optimal algorithm that achieves the exact lower bound on regret for bandits with Gaussian distributions with known variance. Additionally, we provide a finite-sample analysis of CRIMED's regret performance. Notably, CRIMED can effectively handle corruptions with $\varepsilon$ values as high as $\frac{1}{2}$. Furthermore, we develop a tight concentration result for medians in the presence of arbitrary corruptions, even with $\varepsilon$ values up to $\frac{1}{2}$, which may be of independent interest. We also discuss an extension of the algorithm for handling misspecification in Gaussian model.
- Europe > France > Hauts-de-France > Nord > Lille (0.04)
- Asia > Afghanistan > Parwan Province > Charikar (0.04)
- North America > United States > New Jersey > Hudson County > Hoboken (0.04)
- (4 more...)
A Non-Intrusive Correction Algorithm for Classification Problems with Corrupted Data
Hou, Jun, Qin, Tong, Wu, Kailiang, Xiu, Dongbin
A novel correction algorithm is proposed for multi-class classification problems with corrupted training data. The algorithm is non-intrusive, in the sense that it post-processes a trained classification model by adding a correction procedure to the model prediction. The correction procedure can be coupled with any approximators, such as logistic regression, neural networks of various architectures, etc. When training dataset is sufficiently large, we prove that the corrected models deliver correct classification results as if there is no corruption in the training data. For datasets of finite size, the corrected models produce significantly better recovery results, compared to the models without the correction algorithm. All of the theoretical findings in the paper are verified by our numerical examples.
Denoising Criterion for Variational Auto-Encoding Framework
Im, Daniel Im Jiwoong (University of Montreal) | Ahn, Sungjin (University of Montreal) | Memisevic, Roland (University of Montreal) | Bengio, Yoshua (University of Montreal)
Denoising autoencoders (DAE) are trained to reconstruct their clean inputs with noise injected at the input level, while variational autoencoders (VAE) are trained with noise injected in their stochastic hidden layer, with a regularizer that encourages this noise injection. In this paper, we show that injecting noise both in input and in the stochastic hidden layer can be advantageous and we propose a modified variational lower bound as an improved objective function in this setup. When input is corrupted, then the standard VAE lower bound involves marginalizing the encoder conditional distribution over the input noise, which makes the training criterion intractable. Instead, we propose a modified training criterion which corresponds to a tractable bound when input is corrupted. Experimentally, we find that the proposed denoising variational autoencoder (DVAE) yields better average log-likelihood than the VAE and the importance weighted autoencoder on the MNIST and Frey Face datasets.
- North America > Canada > Quebec > Montreal (0.14)
- North America > Canada > Ontario > Toronto (0.14)
- Asia > Middle East > Jordan (0.04)