Prioritizing Samples in Reinforcement Learning with Reducible Loss
–Neural Information Processing Systems
Most reinforcement learning algorithms take advantage of an experience replay buffer to repeatedly train on samples the agent has observed in the past. Not all samples carry the same amount of significance and simply assigning equal importance to each of the samples is a naive strategy. In this paper, we propose a method to prioritize samples based on how much we can learn from a sample.
Neural Information Processing Systems
Feb-11-2026, 08:36:46 GMT
- Country:
- Europe
- North America
- Canada
- British Columbia > Vancouver (0.04)
- Quebec > Montreal (0.04)
- Puerto Rico > San Juan
- San Juan (0.04)
- United States
- Arizona > Maricopa County
- Phoenix (0.04)
- Illinois > Cook County
- Chicago (0.04)
- Arizona > Maricopa County
- Canada
- South America > Brazil
- Pernambuco (0.04)
- Genre:
- Research Report > New Finding (0.46)
- Technology: