G-Safeguard: A Topology-Guided Security Lens and Treatment on LLM-based Multi-agent Systems
Wang, Shilong, Zhang, Guibin, Yu, Miao, Wan, Guancheng, Meng, Fanci, Guo, Chongye, Wang, Kun, Wang, Yang
–arXiv.org Artificial Intelligence
Large Language Model (LLM)-based Multi-agent Systems (MAS) have demonstrated remarkable capabilities in various complex tasks, ranging from collaborative problem-solving to autonomous decision-making. However, as these systems become increasingly integrated into critical applications, their vulnerability to adversarial attacks, misinformation propagation, and unintended behaviors have raised significant concerns. To address this challenge, we introduce G-Safeguard, a topology-guided security lens and treatment for robust LLM-MAS, which leverages graph neural networks to detect anomalies on the multi-agent utterance graph and employ topological intervention for attack remediation. Extensive experiments demonstrate that G-Safeguard: (I) exhibits significant effectiveness under various attack strategies, recovering over 40% of the performance for prompt injection; (II) is highly adaptable to diverse LLM backbones and large-scale MAS; (III) can seamlessly combine with mainstream MAS with security guarantees. The code is available at https://github.com/wslong20/G-safeguard.
arXiv.org Artificial Intelligence
Feb-16-2025
- Country:
- Asia (0.28)
- Genre:
- Research Report (0.50)
- Industry:
- Government > Military (0.87)
- Information Technology > Security & Privacy (1.00)
- Technology: