A Proposal for Evaluating the Operational Risk for ChatBots based on Large Language Models
Pinacho-Davidson, Pedro, Gutierrez, Fernando, Zapata, Pablo, Vergara, Rodolfo, Aqueveque, Pablo
–arXiv.org Artificial Intelligence
The emergence of Generative AI (Gen AI) and Large Language Models (LLMs) has enabled more advanced chatbots capable of human-like interactions. However, these conversational agents introduce a broader set of operational risks that extend beyond traditional cybersecurity considerations. In this work, we propose a novel, instrumented risk-assessment metric that simultaneously evaluates potential threats to three key stakeholders: the service-providing organization, end users, and third parties. Our approach incorporates the technical complexity required to induce erroneous behaviors in the chatbot--ranging from non-induced failures to advanced prompt-injection attacks--as well as contextual factors such as the target industry, user age range, and vulnerability severity. To validate our metric, we leverage Garak, an open-source framework for LLM vulnerability testing. We further enhance Garak to capture a variety of threat vectors (e.g., misinformation, code hallucinations, social engineering, and malicious code generation). Our methodology is demonstrated in a scenario involving chatbots that employ retrieval-augmented generation (RAG), showing how the aggregated risk scores guide both short-term mitigation and longer-term improvements in model design and deployment. The results underscore the importance of multi-dimensional risk assessments in operationalizing secure, reliable AI-driven conversational systems.
arXiv.org Artificial Intelligence
May-9-2025
- Country:
- North America > United States (0.14)
- South America > Chile
- Biobío Region > Concepción Province > Concepción (0.04)
- Genre:
- Research Report > New Finding (0.87)
- Industry:
- Information Technology > Security & Privacy (1.00)
- Technology: