ToolkenGPT: Augmenting Frozen Language Models with Massive Tools via Tool Embeddings Shibo Hao 1
–Neural Information Processing Systems
Augmenting large language models (LLMs) with external tools has emerged as a promising approach to solving complex problems. However, traditional methods, which fine-tune LLMs with tool demonstration data, can be both costly and restricted to a predefined set of tools. Recent in-context learning paradigm alleviates these issues, but the limited context length only allows for a few shots of demonstrations, leading to suboptimal understandings of the tools. Moreover, when there are numerous tools to choose from, in-context learning could completely fail to work. In this paper, we propose an alternative approach, ToolkenGPT, which combines the benefits of both sides.
Neural Information Processing Systems
May-25-2025, 05:00:52 GMT
- Country:
- North America > United States (0.67)
- Genre:
- Overview (0.46)
- Research Report (0.66)
- Workflow (0.46)
- Industry:
- Leisure & Entertainment (1.00)
- Media > Film (0.68)
- Technology: