Self-Taught Self-Correction for Small Language Models
Moskvoretskii, Viktor, Biemann, Chris, Nikishina, Irina
–arXiv.org Artificial Intelligence
Although large language models (LLMs) have achieved remarkable performance across various tasks, they remain prone to errors. A key challenge is enabling them to self-correct. While prior research has relied on external tools or large proprietary models, this work explores self-correction in small language models (SLMs) through iterative fine-tuning using solely self-generated data. We introduce the Self-Taught Self-Correction (STaSC) algorithm, which incorporates multiple algorithmic design choices. Experimental results on a question-answering task demonstrate that STaSC effectively learns self-correction, leading to significant performance improvements. Our analysis further provides insights into the mechanisms of self-correction and the impact of different design choices on learning dynamics and overall performance. To support future research, we release our user-friendly codebase and lightweight models.
arXiv.org Artificial Intelligence
Mar-11-2025
- Country:
- Europe > Middle East
- Malta (0.14)
- North America > United States (0.14)
- Europe > Middle East
- Genre:
- Research Report (0.82)
- Industry:
- Leisure & Entertainment (0.68)
- Media > Film (0.47)
- Technology: