Understanding Model Ensemble in Transferable Adversarial Attack
Yao, Wei, Zhang, Zeliang, Tang, Huayi, Liu, Yong
–arXiv.org Artificial Intelligence
Model ensemble adversarial attack has become a powerful method for generating transferable adversarial examples that can target even unknown models, but its theoretical foundation remains underexplored. To address this gap, we provide early theoretical insights that serve as a roadmap for advancing model ensemble adversarial attack. We first define transferability error to measure the error in adversarial transferability, alongside concepts of diversity and empirical model ensemble Rademacher complexity. We then decompose the transferability error into vulnerability, diversity, and a constant, which rigidly explains the origin of transferability error in model ensemble attack: the vulnerability of an adversarial example to ensemble components, and the diversity of ensemble components. Furthermore, we apply the latest mathematical tools in information theory to bound the transferability error using complexity and generalization terms, contributing to three practical guidelines for reducing transferability error: (1) incorporating more surrogate models, (2) increasing their diversity, and (3) reducing their complexity in cases of overfitting.
arXiv.org Artificial Intelligence
Oct-9-2024
- Country:
- Europe > United Kingdom > England (0.14)
- Genre:
- Research Report (0.81)
- Industry:
- Government > Military (0.92)
- Information Technology > Security & Privacy (1.00)
- Technology: