What Was Your Prompt? A Remote Keylogging Attack on AI Assistants
Weiss, Roy, Ayzenshteyn, Daniel, Amit, Guy, Mirsky, Yisroel
–arXiv.org Artificial Intelligence
AI assistants are becoming an integral part of society, used for asking advice or help in personal and confidential issues. In this paper, we unveil a novel side-channel that can be used to read encrypted responses from AI Assistants over the web: the token-length side-channel. We found that many vendors, including OpenAI and Microsoft, have this side-channel. However, inferring the content of a response from a token-length sequence alone proves challenging. This is because tokens are akin to words, and responses can be several sentences long leading to millions of grammatically correct sentences. In this paper, we show how this can be overcome by (1) utilizing the power of a large language model (LLM) to translate these sequences, (2) providing the LLM with inter-sentence context to narrow the search space and (3) performing a known-plaintext attack by fine-tuning the model on the target model's writing style. Using these methods, we were able to accurately reconstruct 29\% of an AI assistant's responses and successfully infer the topic from 55\% of them. To demonstrate the threat, we performed the attack on OpenAI's ChatGPT-4 and Microsoft's Copilot on both browser and API traffic.
arXiv.org Artificial Intelligence
Mar-14-2024
- Country:
- Asia
- China > Hong Kong (0.04)
- Middle East > Israel (0.14)
- Europe
- North America > United States
- Arkansas > Benton County
- Rogers (0.04)
- California > San Francisco County
- San Francisco (0.04)
- Arkansas > Benton County
- Asia
- Genre:
- Research Report > New Finding (0.46)
- Industry:
- Education (0.93)
- Government (0.67)
- Health & Medicine > Therapeutic Area
- Psychiatry/Psychology (0.46)
- Information Technology > Security & Privacy (1.00)
- Technology: