Exploring Potential Prompt Injection Attacks in Federated Military LLMs and Their Mitigation
Lee, Youngjoon, Park, Taehyun, Lee, Yunho, Gong, Jinu, Kang, Joonhyuk
–arXiv.org Artificial Intelligence
Federated Learning (FL) is increasingly being adopted in military collaborations to develop Large Language Models (LLMs) while preserving data sovereignty. However, prompt injection attacks-malicious manipulations of input prompts-pose new threats that may undermine operational security, disrupt decision-making, and erode trust among allies. This perspective paper highlights four potential vulnerabilities in federated military LLMs: secret data leakage, free-rider exploitation, system disruption, and misinformation spread. To address these potential risks, we propose a human-AI collaborative framework that introduces both technical and policy countermeasures. On the technical side, our framework uses red/blue team wargaming and quality assurance to detect and mitigate adversarial behaviors of shared LLM weights. On the policy side, it promotes joint AI-human policy development and verification of security protocols. Our findings will guide future research and emphasize proactive strategies for emerging military contexts.
arXiv.org Artificial Intelligence
Jan-30-2025
- Country:
- Asia
- Singapore (0.04)
- South Korea
- Europe
- Denmark > Capital Region
- Copenhagen (0.04)
- Italy (0.04)
- Portugal > Porto
- Porto (0.04)
- United Kingdom > England
- Greater London > London (0.04)
- Denmark > Capital Region
- North America
- Canada > British Columbia
- United States
- Florida
- Broward County > Fort Lauderdale (0.04)
- Hillsborough County > University (0.04)
- Massachusetts > Middlesex County
- Cambridge (0.04)
- New York (0.04)
- Pennsylvania > Philadelphia County
- Philadelphia (0.04)
- Texas > Travis County
- Austin (0.14)
- Florida
- Asia
- Genre:
- Research Report > New Finding (0.34)
- Industry:
- Technology: