Guardian of the Ensembles: Introducing Pairwise Adversarially Robust Loss for Resisting Adversarial Attacks in DNN Ensembles
Shukla, Shubhi, Dalui, Subhadeep, Alam, Manaar, Datta, Shubhajit, Mondal, Arijit, Mukhopadhyay, Debdeep, Chakrabarti, Partha Pratim
–arXiv.org Artificial Intelligence
Adversarial attacks rely on transferability, where an adversarial example (AE) crafted on a surrogate classifier tends to mislead a target classifier. Recent ensemble methods demonstrate that AEs are less likely to mislead multiple classifiers in an ensemble. This paper proposes a new ensemble training using a Pairwise Adversarially Robust Loss (PARL) that by construction produces an ensemble of classifiers with diverse decision boundaries. PARL utilizes outputs and gradients of each layer with respect to network parameters in every classifier within the ensemble simultaneously. PARL is demonstrated to achieve higher robustness against black-box transfer attacks than previous ensemble methods as well as adversarial training without adversely affecting clean example accuracy. Extensive experiments using standard Resnet20, WideResnet28-10 classifiers demonstrate the robustness of PARL against state-of-the-art adversarial attacks. While maintaining similar clean accuracy and lesser training time, the proposed architecture has a 24.8% increase in robust accuracy ($\epsilon$ = 0.07) from the state-of-the art method.
arXiv.org Artificial Intelligence
Dec-3-2024
- Country:
- Asia > Middle East
- UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- Europe (0.67)
- North America > United States
- California > Los Angeles County > Long Beach (0.14)
- Asia > Middle East
- Genre:
- Research Report > Promising Solution (0.34)
- Industry:
- Government > Military (0.82)
- Information Technology > Security & Privacy (1.00)
- Technology: