UDAPDR: Unsupervised Domain Adaptation via LLM Prompting and Distillation of Rerankers
Saad-Falcon, Jon, Khattab, Omar, Santhanam, Keshav, Florian, Radu, Franz, Martin, Roukos, Salim, Sil, Avirup, Sultan, Md Arafat, Potts, Christopher
–arXiv.org Artificial Intelligence
Many information retrieval tasks require large labeled datasets for fine-tuning. However, such datasets are often unavailable, and their utility for real-world applications can diminish quickly due to domain shifts. To address this challenge, we develop and motivate a method for using large language models (LLMs) to generate large numbers of synthetic queries cheaply. The method begins by generating a small number of synthetic queries using an expensive LLM. After that, a much less expensive one is used to create large numbers of synthetic queries, which are used to fine-tune a family of reranker models. These rerankers are then distilled into a single efficient retriever for use in the target domain. We show that this technique Figure 1: Overview of UDAPDR. An expensive LLM boosts zero-shot accuracy in long-tail domains like GPT-3 is used to create an initial set of synthetic and achieves substantially lower latency than queries. These are incorporated into a set of prompts for standard reranking methods.
arXiv.org Artificial Intelligence
Oct-13-2023
- Country:
- North America > United States > Minnesota (0.28)
- Genre:
- Research Report > New Finding (0.46)
- Technology: