LLM-Oriented Retrieval Tuner
Sun, Si, Zhang, Hanqing, Liu, Zhiyuan, Bao, Jie, Song, Dawei
–arXiv.org Artificial Intelligence
Dense Retrieval (DR) is now considered as a promising tool to enhance the memorization capacity of Large Language Models (LLM) such as GPT3 and GPT-4 by incorporating external memories. However, due to the paradigm discrepancy between text generation of LLM and DR, it is still an open challenge to integrate the retrieval and generation tasks in a shared LLM. In this paper, we propose an efficient LLM-Oriented Retrieval Tuner, namely LMORT, which decouples DR capacity from base LLM and non-invasively coordinates the optimally aligned and uniform layers of the LLM towards a unified DR space, achieving an efficient and effective DR without tuning the LLM itself. The extensive experiments on six BEIR datasets show that our approach could achieve competitive zero-shot retrieval performance compared to a range of strong DR models while maintaining the generation ability of LLM.
arXiv.org Artificial Intelligence
Mar-4-2024
- Country:
- Asia
- Europe
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Spain > Catalonia
- Barcelona Province > Barcelona (0.04)
- Ireland > Leinster
- North America > United States
- New York > New York County
- New York City (0.04)
- Washington > King County
- Seattle (0.04)
- New York > New York County
- Oceania > Australia
- Genre:
- Research Report (0.82)
- Technology: