Know Your Limits: A Survey of Abstention in Large Language Models

Wen, Bingbing, Yao, Jihan, Feng, Shangbin, Xu, Chenjun, Tsvetkov, Yulia, Howe, Bill, Wang, Lucy Lu

arXiv.org Artificial Intelligence 

But questions of Large language models (LLMs) have demonstrated human values and the answerability of the query generalization capabilities across NLP tasks such itself are difficult to model in terms of model confidence as question answering (QA) (Wei et al., 2022; (Yang et al., 2023). Chowdhery et al., 2022), abstractive summarization (Zhang et al., 2023a), and dialogue generation While prior work demonstrates the potential of (Yi et al., 2024). But these models are also unreliable, abstention in enhancing model safety and reliability having a tendency to "hallucinate" false information (Varshney et al., 2023; Wang et al., 2024c; in their responses (Ji et al., 2023b), generate Zhang et al., 2024a), the study of abstention has overly certain or authoritative responses (Zhou also been constrained to specific QA tasks. This et al., 2024b), answer with incomplete information task-specific approach limits the broader applicability (Zhou et al., 2023b), or produce harmful or of abstention strategies across the diverse dangerous responses (Anwar et al., 2024). In these range of scenarios encountered by general-purpose situations, the model should ideally abstain: to chatbots engaging in open-domain interactions.