Training-Time Action Conditioning for Efficient Real-Time Chunking
Black, Kevin, Ren, Allen Z., Equi, Michael, Levine, Sergey
–arXiv.org Artificial Intelligence
Real-time chunking (RTC) enables vision-language-action models (VLAs) to generate smooth, reactive robot trajectories by asynchronously predicting action chunks and conditioning on previously committed actions via inference-time inpainting. However, this inpainting method introduces computational overhead that increases inference latency. In this work, we propose a simple alternative: simulating inference delay at training time and conditioning on action prefixes directly, eliminating any inference-time overhead. Our method requires no modifications to the model architecture or robot runtime, and can be implemented with only a few additional lines of code. In simulated experiments, we find that training-time RTC outperforms inference-time RTC at higher inference delays. In real-world experiments on box building and espresso making tasks with the $π_{0.6}$ VLA, we demonstrate that training-time RTC maintains both task performance and speed parity with inference-time RTC while being computationally cheaper. Our results suggest that training-time action conditioning is a practical drop-in replacement for inference-time inpainting in real-time robot control.
arXiv.org Artificial Intelligence
Dec-10-2025
- Country:
- Asia > Japan
- Honshū > Tōhoku > Miyagi Prefecture > Sendai (0.04)
- North America > Montserrat (0.04)
- Asia > Japan
- Genre:
- Research Report > New Finding (0.87)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning (1.00)
- Robots (1.00)
- Information Technology > Artificial Intelligence