a94a8800a4b0af45600bab91164849df-Supplemental-Conference.pdf
–Neural Information Processing Systems
Supplementary Material: Can Adversarial Training Be Manipulated By Non-Robust Features? In this part, we discuss several independent (or concurrent) works that are closely related to this work. They also conclude that conventional adversarial training will prevent a drop in accuracy measured both on clean images and adversarial images. In contrast, we focus on a more realistic setting that does not require a larger attack budget. From this perspective, our work is complementary to theirs. This makes the threat of stability attacks more insidious than that of Fu et al. [19].
Neural Information Processing Systems
Aug-17-2025, 12:22:43 GMT