Evaluating Uncertainty Quantification Methods in Argumentative Large Language Models
Zhou, Kevin, Dejl, Adam, Freedman, Gabriel, Chen, Lihu, Rago, Antonio, Toni, Francesca
–arXiv.org Artificial Intelligence
Research in uncertainty quantification (UQ) for large language models (LLMs) is increasingly important towards guaranteeing the reliability of this groundbreaking technology. We explore the integration of LLM UQ methods in argumentative LLMs (ArgLLMs), an explainable LLM framework for decision-making based on computational argumentation in which UQ plays a critical role. We conduct experiments to evaluate ArgLLMs' performance on claim verification tasks when using different LLM UQ methods, inherently performing an assessment of the UQ methods' effectiveness. Moreover, the experimental procedure itself is a novel way of evaluating the effectiveness of UQ methods, especially when intricate and potentially contentious statements are present. Our results demonstrate that, despite its simplicity, direct prompting is an effective UQ strategy in ArgLLMs, outperforming considerably more complex approaches.
arXiv.org Artificial Intelligence
Oct-6-2025
- Country:
- Africa
- Rwanda > Kigali
- Kigali (0.04)
- South Africa (0.04)
- Rwanda > Kigali
- Asia > Singapore (0.05)
- Europe
- Austria (0.04)
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- North America
- Mexico > Mexico City
- Mexico City (0.04)
- United States
- Florida > Miami-Dade County
- Miami (0.14)
- Louisiana > Orleans Parish
- New Orleans (0.04)
- Florida > Miami-Dade County
- Mexico > Mexico City
- Africa
- Genre:
- Research Report > New Finding (1.00)
- Technology: