Instruction Tuning Chronologically Consistent Language Models
He, Songrun, Lv, Linying, Manela, Asaf, Wu, Jimmy
–arXiv.org Artificial Intelligence
We introduce a family of chronologically consistent, instruction-tuned large language models to eliminate lookahead bias. Each model is trained only on data available before a clearly defined knowledge-cutoff date, ensuring strict temporal separation from any post-cutoff data. The resulting framework offers (i) a simple, conversational chat interface, (ii) fully open, fixed model weights that guarantee replicability, and (iii) a conservative lower bound on forecast accuracy, isolating the share of predictability that survives once training leakage is removed. Together, these features provide researchers with an easy-to-use generative AI tool useful for a wide range of prediction tasks that is free of lookahead bias.
arXiv.org Artificial Intelligence
Nov-18-2025
- Country:
- Europe > United Kingdom (0.04)
- North America > United States (1.00)
- Genre:
- Research Report > New Finding (0.46)
- Industry:
- Technology: