KBLaM: Knowledge Base augmented Language Model
Wang, Xi, Mikaelyan, Liana, Isazawa, Taketomo, Hensman, James
–arXiv.org Artificial Intelligence
In this paper, we propose Knowledge Base augmented Language Model (KBLaM), a new method for augmenting Large Language Models (LLMs) with external knowledge. KBLaM works with a knowledge base (KB) constructed from a corpus of documents, transforming each piece of knowledge in the KB into continuous key-value vector pairs via pre-trained sentence encoders with linear adapters and integrating them into pre-trained LLMs via a specialized rectangular attention mechanism. Unlike Retrieval-Augmented Generation, KBLaM eliminates external retrieval modules, and unlike in-context learning, its computational overhead scales linearly with KB size rather than quadratically. Our approach enables integrating a large KB of more than 10K triples into an 8B pre-trained LLM of only 8K context window on one single A100 80GB GPU and allows for dynamic updates without model fine-tuning or retraining. Experiments demonstrate KBLaM's effectiveness in various tasks, including question-answering and open-ended reasoning, while providing interpretable insights into its use of the augmented knowledge.
arXiv.org Artificial Intelligence
Oct-14-2024
- Country:
- North America > United States > California (0.28)
- Genre:
- Personal > Interview (0.68)
- Research Report > New Finding (0.46)
- Industry:
- Banking & Finance (1.00)
- Education (1.00)
- Energy
- Oil & Gas (0.68)
- Renewable > Ocean Energy (0.46)
- Health & Medicine > Consumer Health (1.00)
- Information Technology > Security & Privacy (1.00)
- Leisure & Entertainment (1.00)
- Media > Music (1.00)
- Technology: