STEPER: Step-wise Knowledge Distillation for Enhancing Reasoning Ability in Multi-Step Retrieval-Augmented Language Models
Lee, Kyumin, Jeon, Minjin, Jang, Sanghwan, Yu, Hwanjo
–arXiv.org Artificial Intelligence
Answering complex real-world questions requires step-by-step retrieval and integration of relevant information to generate well-grounded responses. However, existing knowledge distillation methods overlook the need for different reasoning abilities at different steps, hindering transfer in multi-step retrieval-augmented frameworks. To address this, we propose Stepwise Knowledge Distillation for Enhancing Reasoning Ability in Multi-Step Retrieval-Augmented Language Models (StepER). StepER employs step-wise supervision to align with evolving information and reasoning demands across stages. Additionally, it incorporates difficulty-aware training to progressively optimize learning by prioritizing suitable steps. Our method is adaptable to various multi-step retrieval-augmented language models, including those that use retrieval queries for reasoning paths or decomposed questions. Extensive experiments show that StepER outperforms prior methods on multi-hop QA benchmarks, with an 8B model achieving performance comparable to a 70B teacher model.
arXiv.org Artificial Intelligence
Oct-10-2025
- Country:
- Asia
- China > Hong Kong (0.04)
- Middle East
- Pakistan (0.04)
- Russia > Far Eastern Federal District
- Khabarovsk Krai > Khabarovsk (0.04)
- Singapore (0.04)
- Europe
- Belgium > Brussels-Capital Region
- Brussels (0.04)
- France (0.04)
- North Macedonia (0.04)
- Russia (0.04)
- Spain > Catalonia
- Barcelona Province > Barcelona (0.04)
- United Kingdom (0.04)
- Belgium > Brussels-Capital Region
- North America > United States
- Georgia
- Chatham County > Savannah (0.14)
- Fulton County > Atlanta (0.14)
- Nevada (0.04)
- Georgia
- Oceania > Australia
- Queensland (0.04)
- Asia
- Genre:
- Research Report > New Finding (0.93)
- Industry:
- Health & Medicine (0.68)
- Leisure & Entertainment (1.00)
- Media
- Technology: