Encryption-Friendly LLM Architecture
Rho, Donghwan, Kim, Taeseong, Park, Minje, Kim, Jung Woo, Chae, Hyunsik, Cheon, Jung Hee, Ryu, Ernest K.
–arXiv.org Artificial Intelligence
Large language models (LLMs) offer personalized responses based on user interactions, but this use case raises serious privacy concerns. Homomorphic encryption (HE) is a cryptographic protocol supporting arithmetic computations in encrypted states and provides a potential solution for privacy-preserving machine learning (PPML). However, the computational intensity of transformers poses challenges for applying HE to LLMs. In this work, we propose a modified HEfriendly transformer architecture with an emphasis on inference following personalized (private) fine-tuning. Utilizing LoRA fine-tuning and Gaussian kernels, we achieve significant computational speedups--6.94 Our findings provide a viable proof of concept for offering privacy-preserving LLM services in areas where data protection is crucial. One of the many capabilities of LLMs that has received much attention is their ability to provide personalized responses based on user interactions, especially with the use of fine-tuning. However, this use case raises serious concerns about user privacy. In response, regulations such as the GDPR (European Union, 2016) and CCPA (State of California, 2018) have been amended. In Italy, ChatGPT was even temporarily banned (McCallum, 2023), and several major corporations, including Apple and Samsung, have restricted its use within their companies (Mok, 2023). Privacy-preserving machine learning (PPML) refers to methods that use machine learning while protecting data privacy. Techniques for PPML include secure multi-party computation (MPC) (Yao, 1982), differential privacy (Dwork, 2006), and homomorphic encryption (HE) (Rivest et al., 1978; Gentry, 2009). Among these, only MPC and HE offer provable security based on cryptographic assumptions. MPC utilizes communications between parties, but these communications can make it challenging to accelerate and parallelize the heavy computation of neural networks. In contrast, HE supports arithmetic computations in encrypted states without requiring communications.
arXiv.org Artificial Intelligence
Oct-3-2024
- Country:
- Asia
- China > Jiangsu Province
- Yancheng (0.04)
- South Korea > Seoul
- Seoul (0.04)
- China > Jiangsu Province
- Europe
- Italy (0.24)
- Ukraine > Kharkiv Oblast
- Kharkiv (0.04)
- North America > United States
- California (0.34)
- Asia
- Genre:
- Research Report > New Finding (0.66)
- Industry:
- Technology: