Learning to Reject with a Fixed Predictor: Application to Decontextualization
Mohri, Christopher, Andor, Daniel, Choi, Eunsol, Collins, Michael
–arXiv.org Artificial Intelligence
Large language models, often trained with billions of parameters, have achieved impressive performance in recent years (Raffel et al., 2019) and are used in a wide variety of natural language generation tasks. However, their output is sometimes undesirable, with hallucinated content (Maynez et al., 2020; Filippova, 2020), and much work remains to fully understand their properties. In many applications, such as healthcare, question-answering systems, or customer service, incorrect predictions are particularly costly and must be avoided. This motivates the design of algorithms for large language models and other NLP tasks that achieve high precision on a large fraction of the input set, while abstaining on the rest. How can we devise such accurate models that allow a reject option?
arXiv.org Artificial Intelligence
Jan-31-2023
- Country:
- Europe > Spain (0.28)
- North America > United States (0.46)
- Genre:
- Research Report > New Finding (0.68)
- Technology: