Better Membership Inference Privacy Measurement through Discrepancy
Wu, Ruihan, Huang, Pengrun, Chaudhuri, Kamalika
–arXiv.org Artificial Intelligence
Membership Inference Attacks have emerged as a dominant method for empirically measuring privacy leakage from machine learning models. Here, privacy is measured by the {\em{advantage}} or gap between a score or a function computed on the training and the test data. A major barrier to the practical deployment of these attacks is that they do not scale to large well-generalized models -- either the advantage is relatively low, or the attack involves training multiple models which is highly compute-intensive. In this work, inspired by discrepancy theory, we propose a new empirical privacy metric that is an upper bound on the advantage of a family of membership inference attacks. We show that this metric does not involve training multiple models, can be applied to large Imagenet classification models in-the-wild, and has higher advantage than existing metrics on models trained with more recent and sophisticated training recipes. Motivated by our empirical results, we also propose new membership inference attacks tailored to these training losses.
arXiv.org Artificial Intelligence
May-23-2024
- Country:
- North America > United States
- California > San Diego County
- San Diego (0.04)
- New York > New York County
- New York City (0.04)
- Texas (0.05)
- California > San Diego County
- North America > United States
- Genre:
- Research Report > New Finding (0.46)
- Industry:
- Information Technology > Security & Privacy (1.00)
- Technology: