Real-Time Diffusion Policies for Games: Enhancing Consistency Policies with Q-Ensembles
Zhang, Ruoqi, Luo, Ziwei, Sjölund, Jens, Mattsson, Per, Gisslén, Linus, Sestini, Alessandro
–arXiv.org Artificial Intelligence
Diffusion models have shown impressive performance in capturing complex and multi-modal action distributions for game agents, but their slow inference speed prevents practical deployment in real-time game environments. While consistency models offer a promising approach for one-step generation, they often suffer from training instability and performance degradation when applied to policy learning. In this paper, we present CPQE (Consistency Policy with Q-Ensembles), which combines consistency models with Q-ensembles to address these challenges.CPQE leverages uncertainty estimation through Q-ensembles to provide more reliable value function approximations, resulting in better training stability and improved performance compared to classic double Q-network methods. Our extensive experiments across multiple game scenarios demonstrate that CPQE achieves inference speeds of up to 60 Hz -- a significant improvement over state-of-the-art diffusion policies that operate at only 20 Hz -- while maintaining comparable performance to multi-step diffusion approaches. CPQE consistently outperforms state-of-the-art consistency model approaches, showing both higher rewards and enhanced training stability throughout the learning process. These results indicate that CPQE offers a practical solution for deploying diffusion-based policies in games and other real-time applications where both multi-modal behavior modeling and rapid inference are critical requirements.
arXiv.org Artificial Intelligence
Mar-21-2025
- Genre:
- Research Report > New Finding (0.93)
- Industry:
- Information Technology (1.00)
- Leisure & Entertainment > Games
- Computer Games (1.00)
- Technology: