Adversarially Robust Classification based on GLRT
Puranik, Bhagyashree, Madhow, Upamanyu, Pedarsani, Ramtin
Machine learning models are vulnerable to adversarial attacks that can often cause misclassification by introducing small but well designed perturbations. In this paper, we explore, in the setting of classical composite hypothesis testing, a defense strategy based on the generalized likelihood ratio test (GLRT), which jointly estimates the class of interest and the adversarial perturbation. We evaluate the GLRT approach for the special case of binary hypothesis testing in white Gaussian noise under $\ell_{\infty}$ norm-bounded adversarial perturbations, a setting for which a minimax strategy optimizing for the worst-case attack is known. We show that the GLRT approach yields performance competitive with that of the minimax approach under the worst-case attack, and observe that it yields a better robustness-accuracy trade-off under weaker attacks, depending on the values of signal components relative to the attack budget. We also observe that the GLRT defense generalizes naturally to more complex models for which optimal minimax classifiers are not known.
Nov-16-2020
- Country:
- Europe > Sweden
- North America
- Canada
- Alberta > Census Division No. 15
- Improvement District No. 9 > Banff (0.04)
- British Columbia > Metro Vancouver Regional District
- Vancouver (0.05)
- Quebec > Montreal (0.05)
- Alberta > Census Division No. 15
- United States > California
- Santa Barbara County > Santa Barbara (0.04)
- Santa Clara County > San Jose (0.04)
- Canada
- Oceania > Australia
- New South Wales > Sydney (0.04)
- Genre:
- Research Report (0.40)
- Industry:
- Government > Military (0.35)
- Information Technology > Security & Privacy (0.35)
- Technology: