EAD: Elastic-Net Attacks to Deep Neural Networks via Adversarial Examples
Chen, Pin-Yu (IBM Research AI) | Sharma, Yash (The Cooper Union, New York) | Zhang, Huan (University of California, Davis) | Yi, Jinfeng (Tencent AI Lab) | Hsieh, Cho-Jui (University of California, Davis)
Recent studies have highlighted the vulnerability of deep neural networks (DNNs) to adversarial examples — a visually indistinguishable adversarial image can easily be crafted to cause a well-trained model to misclassify. Existing methods for crafting adversarial examples are based on L 2 and L ∞ distortion metrics. However, despite the fact that L 1 distortion accounts for the total variation and encourages sparsity in the perturbation, little has been developed for crafting L 1 -based adversarial examples. In this paper, we formulate the process of attacking DNNs via adversarial examples as an elastic-net regularized optimization problem. Our elastic-net attacks to DNNs (EAD) feature L 1 -oriented adversarial examples and include the state-of-the-art L 2 attack as a special case. Experimental results on MNIST, CIFAR10 and ImageNet show that EAD can yield a distinct set of adversarial examples with small L 1 distortion and attains similar attack performance to the state-of-the-art methods in different attack scenarios. More importantly, EAD leads to improved attack transferability and complements adversarial training for DNNs, suggesting novel insights on leveraging L 1 distortion in adversarial machine learning and security implications of DNNs.
Feb-8-2018
- Country:
- North America > United States > California > Yolo County > Davis (0.14)
- Genre:
- Research Report (1.00)
- Industry:
- Information Technology > Security & Privacy (0.88)
- Technology: