Exploiting Verified Neural Networks via Floating Point Numerical Error
We show how to construct adversarial examples for neural networks with exactly verified robustness against $\ell_{\infty}$-bounded input perturbations by exploiting floating point error. We argue that any exact verification of real-valued neural networks must accurately model the implementation details of any floating point arithmetic used during inference or verification.
Mar-5-2020
- Country:
- Genre:
- Research Report (0.82)
- Industry:
- Information Technology (0.70)
- Technology: