Teaching Language Model Agents How to Self-Improve

Neural Information Processing Systems 

A central piece in enabling intelligent agentic behavior in foundation models is to make them capable of introspecting upon their behavior, reasoning, and correcting their mistakes as more computation or interaction is available. Even the strongest proprietary large language models (LLMs) do not quite exhibit the ability of continually improving their responses sequentially. In this paper, we develop RISE: R ecursive I ntro S p E ction, an approach for fine-tuning LLMs to introduce this capability, despite prior work hypothesizing that this capability may not be possible to attain.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found