Incoherence in goal-conditioned autoregressive models
Karwowski, Jacek, Douglas, Raymond
–arXiv.org Artificial Intelligence
We investigate mathematically the notion of incoherence: a structural issue with reinforcement learning policies derived by naive goal-conditioning of autoregressive models. We focus on the process of re-training models on their own actions, that is, fine-tuning offline-learned policies with online RL. We prove that it decreases incoherence and leads to an improvement in return, and we aim to characterize the resulting trajectory of policies. By re-framing standard notions of control-as-inference and soft Q learning, we establish a three-way correspondence with two other ways of understanding the iterative re-training process: as folding the posterior into the reward and, in the deterministic case, as decreasing the temperature parameter; the correspondence has computational content via the training-inference trade-off. Through soft-conditioning generative models, we discuss the link between incoherence and the effective horizon.
arXiv.org Artificial Intelligence
Oct-9-2025
- Country:
- Europe > United Kingdom
- England > Oxfordshire > Oxford (0.04)
- North America > United States
- Georgia > Fulton County
- Atlanta (0.04)
- Illinois > Cook County
- Chicago (0.04)
- New York > New York County
- New York City (0.04)
- Georgia > Fulton County
- Europe > United Kingdom
- Genre:
- Research Report (0.63)
- Industry:
- Leisure & Entertainment > Games (0.93)
- Technology: