Cross-Domain Transferability of Adversarial Perturbations
Naseer, Muhammad Muzammal, Khan, Salman H., Khan, Muhammad Haris, Khan, Fahad Shahbaz, Porikli, Fatih
–Neural Information Processing Systems
Adversarial examples reveal the blind spots of deep neural networks (DNNs) and represent a major concern for security-critical applications. The transferability of adversarial examples makes real-world attacks possible in black-box settings, where the attacker is forbidden to access the internal parameters of the model. The underlying assumption in most adversary generation methods, whether learning an instance-specific or an instance-agnostic perturbation, is the direct or indirect reliance on the original domain-specific data distribution. In this work, for the first time, we demonstrate the existence of domain-invariant adversaries, thereby showing common adversarial space among different datasets and models. To this end, we propose a framework capable of launching highly transferable attacks that crafts adversarial patterns to mislead networks trained on wholly different domains.
Neural Information Processing Systems
Mar-19-2020, 02:00:59 GMT
- Technology: