Second-Order Adversarial Attack and Certifiable Robustness
Li, Bai, Chen, Changyou, Wang, Wenlin, Carin, Lawrence
We propose a powerful second-order attack method that outperforms existing attack methods on reducing the accuracy of state-of-the-art defense models based on adversarial training. The effectiveness of our attack method motivates an investigation of provable robustness of a defense model. To this end, we introduce a framework that allows one to obtain a certifiable lower bound on the prediction accuracy against adversarial examples. We conduct experiments to show the effectiveness of our attack method. At the same time, our defense models obtain higher accuracies compared to previous works under our proposed attack.
Sep-9-2018
- Genre:
- Research Report > New Finding (0.46)
- Industry:
- Government > Military (0.51)
- Information Technology > Security & Privacy (0.83)
- Technology: