Scaling Goal-based Exploration via Pruning Proto-goals
Bagaria, Akhil, Jiang, Ray, Kumar, Ramana, Schaul, Tom
–arXiv.org Artificial Intelligence
One of the gnarliest challenges in reinforcement learning (RL) is exploration that scales to vast domains, where novelty-, or coverage-seeking behaviour falls short. Goal-directed, purposeful behaviours are able to overcome this, but rely on a good goal space. The core challenge in goal discovery is finding the right balance between generality (not hand-crafted) and tractability (useful, not too many). Our approach explicitly seeks the middle ground, enabling the human designer to specify a vast but meaningful proto-goal space, and an autonomous discovery process to refine this to a narrower space of controllable, reachable, novel, and relevant goals. The effectiveness of goal-conditioned exploration with the latter is then demonstrated in three challenging environments.
arXiv.org Artificial Intelligence
Feb-9-2023
- Country:
- North America > United States > Texas (0.28)
- Genre:
- Research Report (0.50)
- Industry:
- Education (0.46)
- Leisure & Entertainment > Games (0.46)
- Transportation > Passenger (0.47)
- Technology: