RARR: Researching and Revising What Language Models Say, Using Language Models
Gao, Luyu, Dai, Zhuyun, Pasupat, Panupong, Chen, Anthony, Chaganty, Arun Tejasvi, Fan, Yicheng, Zhao, Vincent Y., Lao, Ni, Lee, Hongrae, Juan, Da-Cheng, Guu, Kelvin
–arXiv.org Artificial Intelligence
Language models (LMs) now excel at many tasks such as few-shot learning, question answering, reasoning, and dialog. However, they sometimes generate unsupported or misleading content. A user cannot easily determine whether their outputs are trustworthy or not, because most LMs do not have any built-in mechanism for attribution to external evidence. To enable attribution while still preserving all the powerful advantages of recent generation models, we propose RARR (Retrofit Attribution using Research and Revision), a system that 1) automatically finds attribution for the output of any text generation model and 2) post-edits the output to fix unsupported content while preserving the original output as much as possible. When applied to the output of several state-of-the-art LMs on a diverse set of generation tasks, we find that RARR significantly improves attribution while otherwise preserving the original input to a much greater degree than previously explored edit models. Furthermore, the implementation of RARR requires only a handful of training examples, a large language model, and standard web search.
arXiv.org Artificial Intelligence
May-31-2023
- Country:
- Africa (0.69)
- Asia (0.94)
- Europe > United Kingdom (0.68)
- North America > United States (1.00)
- Genre:
- Research Report > New Finding (0.46)
- Industry:
- Government > Regional Government
- Health & Medicine > Therapeutic Area (0.94)
- Leisure & Entertainment
- Games (0.68)
- Sports > Basketball (0.93)
- Media
- Film (1.00)
- Television (1.00)
- Technology: