Vicious Classifiers: Data Reconstruction Attack at Inference Time
Malekzadeh, Mohammad, Gunduz, Deniz
–arXiv.org Artificial Intelligence
Privacy-preserving inference in edge computing paradigms encourages the users of machine-learning services to locally run a model on their private input, for a target task, and only share the model's outputs with the server. We study how a vicious server can reconstruct the input data by observing only the model's outputs, while keeping the target accuracy very close to that of a honest server: by jointly training a target model (to run at users' side) and an attack model for data reconstruction (to secretly use at server's side). We present a new measure to assess the reconstruction risk in edge inference. Our evaluations on six benchmark datasets demonstrate that the model's input can be approximately reconstructed from the outputs of a single target inference. We propose a potential defense mechanism that helps to distinguish vicious versus honest classifiers at inference time. We discuss open challenges and directions for future studies and release our code as a benchmark for future work.
arXiv.org Artificial Intelligence
Dec-5-2023
- Country:
- Europe > United Kingdom > England
- Cambridgeshire > Cambridge (0.14)
- Greater London > London (0.04)
- Oxfordshire > Oxford (0.04)
- Europe > United Kingdom > England
- Genre:
- Research Report > New Finding (0.46)
- Industry:
- Education (0.68)
- Information Technology > Security & Privacy (1.00)
- Technology: