Goto

Collaborating Authors

 Machiraju, Harshitha


Frequency-Based Vulnerability Analysis of Deep Learning Models against Image Corruptions

arXiv.org Artificial Intelligence

Deep learning models often face challenges when handling real-world image corruptions. In response, researchers have developed image corruption datasets to evaluate the performance of deep neural networks in handling such corruptions. However, these datasets have a significant limitation: they do not account for all corruptions encountered in real-life scenarios. To address this gap, we present MUFIA (Multiplicative Filter Attack), an algorithm designed to identify the specific types of corruptions that can cause models to fail. Our algorithm identifies the combination of image frequency components that render a model susceptible to misclassification while preserving the semantic similarity to the original image. We find that even state-of-the-art models trained to be robust against known common corruptions struggle against the low visibility-based corruptions crafted by MUFIA. This highlights the need for more comprehensive approaches to enhance model robustness against a wider range of real-world image corruptions.


A comment on Guo et al. [arXiv:2206.11228]

arXiv.org Artificial Intelligence

In a recent article, Guo et al. (2022) report that adversarially trained neural representations in deep networks may already be as robust as corresponding primate IT neural representations. By careful perturbation, the authors could change the preferred viewing image of IT neurons---for example, a neuron that had the highest firing rate for images of pressure gauges responded most strongly to images of dogs when minor perturbation to the image was applied. The authors report that the degree of image perturbation required to do this may be even lower than for adversarially trained deep networks. The authors interpret this result by posing an apparent paradox: 'How is it that primate visual perception seems so robust yet its fundamental units of computation are far more sensitive than expected?' While we find the paper's primary experiment illuminating, we have doubts about the interpretation and phrasing of the results presented in the paper.


Bio-inspired Robustness: A Review

arXiv.org Artificial Intelligence

Deep convolutional neural networks (DCNNs) have revolutionized computer vision and are often advocated as good models of the human visual system. However, there are currently many shortcomings of DCNNs, which preclude them as a model of human vision. For example, in the case of adversarial attacks, where adding small amounts of noise to an image, including an object, can lead to strong misclassification of that object. But for humans, the noise is often invisible. If vulnerability to adversarial noise cannot be fixed, DCNNs cannot be taken as serious models of human vision. Many studies have tried to add features of the human visual system to DCNNs to make them robust against adversarial attacks. However, it is not fully clear whether human vision inspired components increase robustness because performance evaluations of these novel components in DCNNs are often inconclusive. We propose a set of criteria for proper evaluation and analyze different models according to these criteria. We finally sketch future efforts to make DCCNs one step closer to the model of human vision.


Harnessing the Vulnerability of Latent Layers in Adversarially Trained Models

arXiv.org Machine Learning

Neural networks are vulnerable to adversarial attacks -- small visually imperceptible crafted noise which when added to the input drastically changes the output. The most effective method of defending against these adversarial attacks is to use the methodology of adversarial training. We analyze the adversarially trained robust models to study their vulnerability against adversarial attacks at the level of the latent layers. Our analysis reveals that contrary to the input layer which is robust to adversarial attack, the latent layer of these robust models are highly susceptible to adversarial perturbations of small magnitude. Leveraging this information, we introduce a new technique Latent Adversarial Training (LAT) which comprises of fine-tuning the adversarially trained models to ensure the robustness at the feature layers. We also propose Latent Attack (LA), a novel algorithm for construction of adversarial examples. LAT results in minor improvement in test accuracy and leads to a state-of-the-art adversarial accuracy against the universal first-order adversarial PGD attack which is shown for the MNIST, CIFAR-10, CIFAR-100 datasets.