Pre-training with Large Language Model-based Document Expansion for Dense Passage Retrieval
Ma, Guangyuan, Wu, Xing, Wang, Peng, Lin, Zijia, Hu, Songlin
–arXiv.org Artificial Intelligence
In this paper, we systematically study the potential of pre-training with Large Language Model(LLM)-based document expansion for dense passage retrieval. Concretely, we leverage the capabilities of LLMs for document expansion, i.e. query generation, and effectively transfer expanded knowledge to retrievers using pre-training strategies tailored for passage retrieval. These strategies include contrastive learning and bottlenecked query generation. Furthermore, we incorporate a curriculum learning strategy to reduce the reliance on LLM inferences. Experimental results demonstrate that pre-training with LLM-based document expansion significantly boosts the retrieval performance on large-scale web-search tasks. Our work shows strong zero-shot and out-of-domain retrieval abilities, making it more widely applicable for retrieval when initializing with no human-labeled data.
arXiv.org Artificial Intelligence
Aug-16-2023
- Country:
- Africa
- Ethiopia > Addis Ababa
- Addis Ababa (0.04)
- Rwanda > Kigali
- Kigali (0.04)
- Ethiopia > Addis Ababa
- Asia
- China > Beijing
- Beijing (0.04)
- Middle East > UAE
- Abu Dhabi Emirate > Abu Dhabi (0.04)
- Singapore (0.04)
- China > Beijing
- Europe
- France
- Provence-Alpes-Côte d'Azur > Bouches-du-Rhône
- Marseille (0.04)
- Île-de-France > Paris
- Paris (0.04)
- Provence-Alpes-Côte d'Azur > Bouches-du-Rhône
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Spain (0.04)
- France
- North America
- Canada
- Dominican Republic (0.04)
- United States
- District of Columbia > Washington (0.04)
- Minnesota > Hennepin County
- Minneapolis (0.04)
- Washington > King County
- Seattle (0.04)
- Africa
- Genre:
- Research Report > New Finding (0.66)
- Technology: