Posterior Sampling with Delayed Feedback for Reinforcement Learning with Linear Function Approximation
–Neural Information Processing Systems
Recent studies in reinforcement learning (RL) have made significant progress by leveraging function approximation to alleviate the sample complexity hurdle for better performance. Despite the success, existing provably efficient algorithms typically rely on the accessibility of immediate feedback upon taking actions. The failure to account for the impact of delay in observations can significantly degrade the performance of real-world systems due to the regret blow-up. In this work, we tackle the challenge of delayed feedback in RL with linear function approximation by employing posterior sampling, which has been shown to empirically outperform the popular UCB algorithms in a wide range of regimes. We first introduce Delayed-PSVI, an optimistic value-based algorithm that effectively explores the value function space via noise perturbation with posterior sampling.
Neural Information Processing Systems
Feb-19-2026, 00:51:01 GMT
- Country:
- Asia > Middle East
- Jordan (0.04)
- North America > United States
- California
- San Diego County > San Diego (0.04)
- Santa Barbara County > Santa Barbara (0.04)
- California
- Asia > Middle East
- Genre:
- Research Report (0.87)
- Industry:
- Technology: