Teaching Pretrained Language Models to Think Deeper with Retrofitted Recurrence
McLeish, Sean, Li, Ang, Kirchenbauer, John, Kalra, Dayal Singh, Bartoldson, Brian R., Kailkhura, Bhavya, Schwarzschild, Avi, Geiping, Jonas, Goldstein, Tom, Goldblum, Micah
–arXiv.org Artificial Intelligence
Recent advances in depth-recurrent language models show that recurrence can decouple train-time compute and parameter count from test-time compute. In this work, we study how to convert existing pretrained non-recurrent language models into depth-recurrent models. We find that using a curriculum of recurrences to increase the effective depth of the model over the course of training preserves performance while reducing total computational cost. In our experiments, on mathematics, we observe that converting pretrained models to recurrent ones results in better performance at a given compute budget than simply post-training the original non-recurrent language model.
arXiv.org Artificial Intelligence
Nov-11-2025
- Country:
- Asia
- Middle East > Jordan (0.04)
- Myanmar > Tanintharyi Region
- Dawei (0.04)
- Europe
- Germany > Baden-Württemberg
- Tübingen Region > Tübingen (0.04)
- Italy > Calabria
- Catanzaro Province > Catanzaro (0.04)
- Germany > Baden-Württemberg
- North America > United States
- Maryland > Prince George's County
- College Park (0.04)
- North Carolina (0.04)
- Maryland > Prince George's County
- Asia
- Genre:
- Research Report > New Finding (0.48)
- Industry:
- Education (0.67)
- Energy (0.68)
- Government > Regional Government
- Technology: