Defending Neural Backdoors via Generative Distribution Modeling
Qiao, Ximing, Yang, Yukun, Li, Hai
–Neural Information Processing Systems
Neural backdoor attack is emerging as a severe security threat to deep learning, while the capability of existing defense methods is limited, especially for complex backdoor triggers. In the work, we explore the space formed by the pixel values of all possible backdoor triggers. An original trigger used by an attacker to build the backdoored model represents only a point in the space. It then will be generalized into a distribution of valid triggers, all of which can influence the backdoored model. Thus, previous methods that model only one point of the trigger distribution is not sufficient. Getting the entire trigger distribution, e.g., via generative modeling, is a key of effective defense.
Neural Information Processing Systems
Mar-19-2020, 02:30:37 GMT