SPICE: Self-Play In Corpus Environments Improves Reasoning
Liu, Bo, Jin, Chuanyang, Kim, Seungone, Yuan, Weizhe, Zhao, Wenting, Kulikov, Ilia, Li, Xian, Sukhbaatar, Sainbayar, Lanchantin, Jack, Weston, Jason
–arXiv.org Artificial Intelligence
Self-improving systems require environmental interaction for continuous adaptation. We introduce SPICE (Self-Play In Corpus Environments), a reinforcement learning framework where a single model acts in two roles: a Challenger that mines documents from a large corpus to generate diverse reasoning tasks, and a Reasoner that solves them. Through adversarial dynamics, the Challenger creates an automatic curriculum at the frontier of the Reasoner's capability, while corpus grounding provides the rich, near-inexhaustible external signal necessary for sustained improvement. Unlike existing ungrounded self-play methods that offer more limited benefits, SPICE achieves consistent gains across mathematical (+8.9%) and general reasoning (+9.8%) benchmarks on multiple model families. Our analysis reveals how document grounding is a key ingredient in SPICE to continuously generate its own increasingly challenging goals and achieve them, enabling sustained self-improvement.
arXiv.org Artificial Intelligence
Oct-29-2025
- Country:
- Asia
- China > Jiangsu Province
- Yancheng (0.04)
- Middle East > Jordan (0.04)
- Singapore (0.04)
- China > Jiangsu Province
- Europe > Italy
- Calabria > Catanzaro Province > Catanzaro (0.04)
- North America > United States
- Virginia (0.04)
- Asia
- Genre:
- Research Report (0.64)
- Industry:
- Leisure & Entertainment > Games (1.00)
- Technology: