Towards Efficient and Scalable Training of Differentially Private Deep Learning
Beltran, Sebastian Rodriguez, Tobaben, Marlon, Loppi, Niki, Honkela, Antti
–arXiv.org Artificial Intelligence
Differentially private stochastic gradient descent (DP-SGD) is the standard algorithm for training machine learning models under differential privacy (DP). The major drawback of DP-SGD is the drop in utility which prior work has comprehensively studied. However, in practice another major drawback that hinders the large-scale deployment is the significantly higher computational cost. We conduct a comprehensive empirical study to quantify the computational cost of training deep learning models under DP and benchmark methods that aim at reducing the cost. Among these are more efficient implementations of DP-SGD and training with lower precision. Finally, we study the scaling behaviour using up to 80 GPUs.
arXiv.org Artificial Intelligence
Jun-25-2024
- Country:
- Europe
- Austria > Vienna (0.04)
- Finland > Uusimaa
- Helsinki (0.05)
- Spain > Canary Islands (0.04)
- United Kingdom > England
- Greater London > London (0.04)
- North America
- Canada > Ontario
- Toronto (0.14)
- United States
- California
- Los Angeles County > Long Beach (0.04)
- San Diego County > La Jolla (0.04)
- San Francisco County > San Francisco (0.04)
- Hawaii > Honolulu County
- Honolulu (0.04)
- Louisiana > Orleans Parish
- New Orleans (0.04)
- New York > New York County
- New York City (0.04)
- Texas
- Harris County > Houston (0.04)
- Travis County > Austin (0.04)
- California
- Canada > Ontario
- Europe
- Genre:
- Research Report > New Finding (0.68)
- Industry:
- Government > Regional Government (0.68)
- Information Technology > Security & Privacy (0.93)
- Technology: