Noise Attention Learning: Enhancing Noise Robustness by Gradient Scaling Yang Bo Department of Computing and Software Department of Computing and Software McMaster University
–Neural Information Processing Systems
Machine learning has been highly successful in data-driven applications but is often hampered when the data contains noise, especially label noise. When trained on noisy labels, deep neural networks tend to fit all noisy labels, resulting in poor generalization. To handle this problem, a common idea is to force the model to fit only clean samples rather than mislabeled ones. In this paper, we propose a simple yet effective method that automatically distinguishes the mislabeled samples and prevents the model from memorizing them, named Noise Attention Learning. In our method, we introduce an attention branch to produce attention weights based on representations of samples.
Neural Information Processing Systems
Mar-27-2025, 07:03:51 GMT