The Measure of Deception: An Analysis of Data Forging in Machine Unlearning
Dixit, Rishabh, Hui, Yuan, Saab, Rayan
Motivated by privacy regulations and the need to mitigate the effects of harmful data, machine unlearning seeks to modify trained models so that they effectively ``forget'' designated data. A key challenge in verifying unlearning is forging -- adversarially crafting data that mimics the gradient of a target point, thereby creating the appearance of unlearning without actually removing information. To capture this phenomenon, we consider the collection of data points whose gradients approximate a target gradient within tolerance $ε$ -- which we call an $ε$-forging set -- and develop a framework for its analysis. For linear regression and one-layer neural networks, we show that the Lebesgue measure of this set is small. It scales on the order of $ε$, and when $ε$ is small enough, $ε^d$. More generally, under mild regularity assumptions, we prove that the forging set measure decays as $ε^{(d-r)/2}$, where $d$ is the data dimension and $r
Sep-9-2025
- Country:
- Europe > Czechia
- Prague (0.04)
- North America > United States
- California > San Diego County > San Diego (0.04)
- Europe > Czechia
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Information Technology > Security & Privacy (1.00)
- Law (1.00)
- Technology:
- Information Technology
- Artificial Intelligence > Machine Learning
- Neural Networks (1.00)
- Statistical Learning
- Gradient Descent (0.46)
- Regression (0.48)
- Data Science (0.87)
- Security & Privacy (1.00)
- Artificial Intelligence > Machine Learning
- Information Technology