Policy Gradient Guidance Enables Test Time Control
Qi, Jianing, Tang, Hao, Zhu, Zhigang
–arXiv.org Artificial Intelligence
We introduce Policy Gradient Guidance (PGG), a simple extension of classifier-free guidance from diffusion models to classical policy gradient methods. PGG augments the policy gradient with an unconditional branch and interpolates conditional and unconditional branches, yielding a test-time control knob that modulates behavior without retraining. We provide a theoretical derivation showing that the additional normalization term vanishes under advantage estimation, leading to a clean guided policy gradient update. Empirically, we evaluate PGG on discrete and continuous control benchmarks. We find that conditioning dropout-central to diffusion guidance-offers gains in simple discrete tasks and low sample regimes, but dropout destabilizes continuous control. Training with modestly larger guidance ($γ>1$) consistently improves stability, sample efficiency, and controllability. Our results show that guidance, previously confined to diffusion policies, can be adapted to standard on-policy methods, opening new directions for controllable online reinforcement learning.
arXiv.org Artificial Intelligence
Oct-3-2025
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe > Portugal
- North America > United States (0.14)
- Asia > Middle East
- Genre:
- Research Report > New Finding (0.86)
- Industry:
- Government (0.46)
- Technology: