On the interplay of Explainability, Privacy and Predictive Performance with Explanation-assisted Model Extraction
Ezzeddine, Fatima, Akel, Rinad, Sbeity, Ihab, Giordano, Silvia, Langheinrich, Marc, Ayoub, Omran
–arXiv.org Artificial Intelligence
Machine Learning as a Service (MLaaS) has gained important attraction as a means for deploying powerful predictive models, offering ease of use that enables organizations to leverage advanced analytics without substantial investments in specialized infrastructure or expertise. However, MLaaS platforms must be safeguarded against security and privacy attacks, such as model extraction (MEA) attacks. The increasing integration of explainable AI (XAI) within MLaaS has introduced an additional privacy challenge, as attackers can exploit model explanations particularly counterfactual explanations (CFs) to facilitate MEA. In this paper, we investigate the trade offs among model performance, privacy, and explainability when employing Differential Privacy (DP), a promising technique for mitigating CF facilitated MEA. We evaluate two distinct DP strategies: implemented during the classification model training and at the explainer during CF generation.
arXiv.org Artificial Intelligence
May-15-2025
- Country:
- Asia > Middle East
- Lebanon > Beirut Governorate
- Beirut (0.04)
- Republic of Türkiye > Istanbul Province
- Istanbul (0.04)
- Lebanon > Beirut Governorate
- Europe
- Middle East > Republic of Türkiye
- Istanbul Province > Istanbul (0.04)
- Switzerland (0.05)
- Middle East > Republic of Türkiye
- North America > United States
- California (0.04)
- Asia > Middle East
- Genre:
- Research Report > Promising Solution (0.48)
- Industry:
- Information Technology > Security & Privacy (1.00)
- Technology: