point maze
ASPiRe: Adaptive Skill Priors for Reinforcement Learning
We find that the sample size has almost no impact on the learning. Notice that the target KL divergence imposes on Ant Maze is higher than the one on Point Maze. "space" to explore around the composite skill prior. As target KL divergence increases, the learned policy will receive less guidance from the prior. The algorithm is not sensitive to this parameter.
Adaptable Hindsight Experience Replay for Search-Based Learning
Vazaios, Alexandros, Brugger, Jannis, Derstroff, Cedric, Kersting, Kristian, Mezini, Mira
AlphaZero-like Monte Carlo Tree Search systems, originally introduced for two-player games, dynamically balance exploration and exploitation using neural network guidance. This combination makes them also suitable for classical search problems. However, the original method of training the network with simulation results is limited in sparse reward settings, especially in the early stages, where the network cannot yet give guidance. Hindsight Experience Replay (HER) addresses this issue by relabeling unsuccessful trajectories from the search tree as supervised learning signals. We introduce Adaptable HER (\ours{}), a flexible framework that integrates HER with AlphaZero, allowing easy adjustments to HER properties such as relabeled goals, policy targets, and trajectory selection. Our experiments, including equation discovery, show that the possibility of modifying HER is beneficial and surpasses the performance of pure supervised or reinforcement learning.
- Europe > Germany > Hesse > Darmstadt Region > Darmstadt (0.06)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- Europe > Norway > Eastern Norway > Oslo (0.04)
- Asia > Thailand > Bangkok > Bangkok (0.04)
- Research Report > Promising Solution (0.34)
- Research Report > New Finding (0.34)
ASPiRe: Adaptive Skill Priors for Reinforcement Learning
We find that the sample size has almost no impact on the learning. Notice that the target KL divergence imposes on Ant Maze is higher than the one on Point Maze. "space" to explore around the composite skill prior. As target KL divergence increases, the learned policy will receive less guidance from the prior. The algorithm is not sensitive to this parameter.
Teachable Reinforcement Learning via Advice Distillation
Watkins, Olivia, Darrell, Trevor, Abbeel, Pieter, Andreas, Jacob, Gupta, Abhishek
Training automated agents to complete complex tasks in interactive environments is challenging: reinforcement learning requires careful hand-engineering of reward functions, imitation learning requires specialized infrastructure and access to a human expert, and learning from intermediate forms of supervision (like binary preferences) is time-consuming and extracts little information from each human intervention. Can we overcome these challenges by building agents that learn from rich, interactive feedback instead? We propose a new supervision paradigm for interactive learning based on "teachable" decision-making systems that learn from structured advice provided by an external teacher. We begin by formalizing a class of human-in-the-loop decision making problems in which multiple forms of teacher-provided advice are available to a learner. We then describe a simple learning algorithm for these problems that first learns to interpret advice, then learns from advice to complete tasks even in the absence of human supervision. In puzzle-solving, navigation, and locomotion domains, we show that agents that learn from advice can acquire new skills with significantly less human supervision than standard reinforcement learning algorithms and often less than imitation learning.
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- Europe > Netherlands > North Brabant > Eindhoven (0.04)
- Asia > Middle East > Jordan (0.04)
ASPiRe:Adaptive Skill Priors for Reinforcement Learning
Xu, Mengda, Veloso, Manuela, Song, Shuran
We introduce ASPiRe (Adaptive Skill Prior for RL), a new approach that leverages prior experience to accelerate reinforcement learning. Unlike existing methods that learn a single skill prior from a large and diverse dataset, our framework learns a library of different distinction skill priors (i.e., behavior priors) from a collection of specialized datasets, and learns how to combine them to solve a new task. This formulation allows the algorithm to acquire a set of specialized skill priors that are more reusable for downstream tasks; however, it also brings up additional challenges of how to effectively combine these unstructured sets of skill priors to form a new prior for new tasks. Specifically, it requires the agent not only to identify which skill prior(s) to use but also how to combine them (either sequentially or concurrently) to form a new prior. To achieve this goal, ASPiRe includes Adaptive Weight Module (AWM) that learns to infer an adaptive weight assignment between different skill priors and uses them to guide policy learning for downstream tasks via weighted Kullback-Leibler divergences. Our experiments demonstrate that ASPiRe can significantly accelerate the learning of new downstream tasks in the presence of multiple priors and show improvement on competitive baselines.
- North America > United States > New York > New York County > New York City (0.04)
- Asia > Middle East > Jordan (0.04)
- Asia > Japan > Honshū > Chūbu > Toyama Prefecture > Toyama (0.04)
- Research Report (1.00)
- Instructional Material > Course Syllabus & Notes (0.34)
- Education (0.46)
- Banking & Finance (0.46)