Goto

Collaborating Authors

 leakage



fa84632d742f2729dc32ce8cb5d49733-Supplemental.pdf

Neural Information Processing Systems

However, a gradient is often insufficient to reconstruct the user data without any prior knowledge. By exploiting a generative model pretrained on the data distribution, we demonstrate that data privacy can be easily breached.



In Differential Privacy, There is Truth: On Vote Leakage in Ensemble Private Learning Jiaqi Wang

Neural Information Processing Systems

When learning from sensitive data, care must be taken to ensure that training algorithms address privacy concerns. The canonical Private Aggregation of Teacher Ensembles, or P A TE, computes output labels by aggregating the predictions of a (possibly distributed) collection of teacher models via a voting mechanism. The mechanism adds noise to attain a differential privacy guarantee with respect to the teachers' training data. In this work, we observe that this use of noise, which makes P A TE predictions stochastic, enables new forms of leakage of sensitive information. For a given input, our adversary exploits this stochasticity to extract high-fidelity histograms of the votes submitted by the underlying teachers. From these histograms, the adversary can learn sensitive attributes of the input such as race, gender, or age. Although this attack does not directly violate the differential privacy guarantee, it clearly violates privacy norms and expectations, and would not be possible at all without the noise inserted to obtain differential privacy. In fact, counter-intuitively, the attack becomes easier as we add more noise to provide stronger differential privacy. We hope this encourages future work to consider privacy holistically rather than treat differential privacy as a panacea.


ProPILE: Probing Privacy Leakage in Large Language Models Siwon Kim 1, Sangdoo Y un 3 Hwaran Lee 3 Martin Gubri

Neural Information Processing Systems

The rapid advancement and widespread use of large language models (LLMs) have raised significant concerns regarding the potential leakage of personally identifiable information (PII). These models are often trained on vast quantities of web-collected data, which may inadvertently include sensitive personal data.





GlanceNets: Interpretable, Leak-proof Concept-basedModels

Neural Information Processing Systems

One reason is that the notion of interpretability is notoriously challenging to pin down, andtherefore existing CBMs rely ondifferent heuristics--such asencouraging theconcepts tobe sparse [1], orthonormal to each other [5], or match the contents of concrete examples [3]--with unclear properties and incompatible goals.