AgentGuard: Runtime Verification of AI Agents
–arXiv.org Artificial Intelligence
The rapid evolution to autonomous, agentic AI systems introduces significant risks due to their inherent unpredictability and emergent behaviors; this also renders traditional verification methods inadequate and necessitates a shift towards probabilistic guarantees where the question is no longer if a system will fail, but the probability of its failure within given constraints. This paper presents AgentGuard, a framework for runtime verification of Agentic AI systems that provides continuous, quantitative assurance through a new paradigm called Dynamic Probabilistic Assurance. AgentGuard operates as an inspection layer that observes an agent's raw I/O and abstracts it into formal events corresponding to transitions in a state model. It then uses online learning to dynamically build and update a Markov Decision Process (MDP) that formally models the agent's emergent behavior. Using probabilistic model checking, the framework then verifies quantitative properties in real-time.
arXiv.org Artificial Intelligence
Sep-30-2025
- Country:
- Europe
- France > Occitanie
- Haute-Garonne > Toulouse (0.04)
- Netherlands > South Holland
- Delft (0.04)
- France > Occitanie
- Europe
- Genre:
- Research Report (0.64)
- Industry:
- Education (0.35)