Understanding the Relationship between Prompts and Response Uncertainty in Large Language Models
Zhang, Ze Yu, Verma, Arun, Doshi-Velez, Finale, Low, Bryan Kian Hsiang
–arXiv.org Artificial Intelligence
Large language models (LLMs) have demonstrated impressive performance across a variety of tasks (Google, 2023; OpenAI, 2023; Zhao et al., 2023). This success has led to their widespread adoption and significant involvement in various decision-making applications, such as healthcare (Karabacak and Margetis, 2023; Sallam, 2023; Yang et al., 2023), education (Xiao et al., 2023), finance (Wu et al., 2023b), and law (Zhang et al., 2023a). However, despite their rapid adoption, the reliability of LLMs in handling high-stakes tasks has yet to be demonstrated (Arkoudas, 2023; Huang et al., 2023a). The reliability is particularly critical in domains such as healthcare, where model responses can have immediate and significant impacts on human behavior and hence their well-being (Ji et al., 2023). Therefore, understanding LLMs' reasoning and decision-making processes and how they influence response uncertainty is critical for their safe and reliable deployment.
arXiv.org Artificial Intelligence
Aug-21-2024
- Country:
- Africa > Democratic Republic of the Congo (0.04)
- Asia
- Europe
- Germany (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Oxfordshire > Oxford (0.04)
- North America > United States (0.67)
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Technology: