From Capabilities to Performance: Evaluating Key Functional Properties of LLM Architectures in Penetration Testing
Huang, Lanxiao, Dave, Daksh, Cody, Tyler, Beling, Peter, Jin, Ming
–arXiv.org Artificial Intelligence
Large language models (LLMs) are increasingly used to automate or augment penetration testing, but their effectiveness and reliability across attack phases remain unclear. We present a comprehensive evaluation of multiple LLM-based agents, from single-agent to modular designs, across realistic penetration testing scenarios, measuring empirical performance and recurring failure patterns. We also isolate the impact of five core functional capabilities via targeted augmentations: Global Context Memory (GCM), Inter-Agent Messaging (IAM), Context-Conditioned Invocation (CCI), Adaptive Planning (AP), and Real-Time Monitoring (RTM). These interventions support, respectively: (i) context coherence and retention, (ii) inter-component coordination and state management, (iii) tool use accuracy and selective execution, (iv) multi-step strategic planning, error detection, and recovery, and (v) real-time dynamic responsiveness. Our results show that while some architectures natively exhibit subsets of these properties, targeted augmentations substantially improve modular agent performance, especially in complex, multi-step, and real-time penetration testing tasks.
arXiv.org Artificial Intelligence
Nov-14-2025
- Country:
- Europe > Slovenia
- Central Slovenia > Municipality of Komenda > Komenda (0.04)
- North America > United States
- Virginia (0.04)
- Europe > Slovenia
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Government > Military
- Cyberwarfare (0.70)
- Information Technology > Security & Privacy (1.00)
- Law (0.92)
- Government > Military
- Technology: