Enhancing Robustness of LLM-Driven Multi-Agent Systems through Randomized Smoothing
Hu, Jinwei, Dong, Yi, Ding, Zhengtao, Huang, Xiaowei
–arXiv.org Artificial Intelligence
This paper presents a defense framework for enhancing the safety of large language model (LLM) empowered multi-agent systems (MAS) in safety-critical domains such as aerospace. We apply randomized smoothing, a statistical robustness certification technique, to the MAS consensus context, enabling probabilistic guarantees on agent decisions under adversarial influence. Unlike traditional verification methods, our approach operates in black-box settings and employs a two-stage adaptive sampling mechanism to balance robustness and computational efficiency. Simulation results demonstrate that our method effectively prevents the propagation of adversarial behaviors and hallucinations while maintaining consensus performance. This work provides a practical and scalable path toward safe deployment of LLM-based MAS in real-world, high-stakes environments.
arXiv.org Artificial Intelligence
Jul-8-2025
- Country:
- Asia > Thailand
- Europe > United Kingdom (0.04)
- North America
- Mexico > Mexico City
- Mexico City (0.04)
- United States (0.04)
- Mexico > Mexico City
- Genre:
- Research Report (1.00)
- Industry:
- Information Technology (1.00)
- Transportation > Air (1.00)
- Technology: