With Friends Like These, Who Needs Adversaries?
Saumya Jetley, Nicholas Lord, Philip Torr
–Neural Information Processing Systems
The vulnerability of deep image classification networks to adversarial attack is now well known, but less well understood. Via a novel experimental analysis, we illustrate some facts about deep convolutional networks for image classification that shed new light on their behaviour and how it connects to the problem of adversaries. In short, the celebrated performance of these networks and their vulnerability to adversarial attack are simply two sides of the same coin: the input image-space directions along which the networks are most vulnerable to attack are the same directions which they use to achieve their classification performance in the first place. We develop this result in two main steps. The first uncovers the fact that classes tend to be associated with specific image-space directions. This is shown by an examination of the class-score outputs of nets as functions of 1D movements along these directions. This provides a novel perspective on the existence of universal adversarial perturbations. The second is a clear demonstration of the tight coupling between classification performance and vulnerability to adversarial attack within the spaces spanned by these directions.
Neural Information Processing Systems
May-26-2025, 07:48:38 GMT
- Country:
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.40)
- Industry:
- Information Technology (0.75)
- Technology: