Learning Reusable Options for Multi-Task Reinforcement Learning
Garcia, Francisco M., Nota, Chris, Thomas, Philip S.
–arXiv.org Artificial Intelligence
One of the main reasons why RL has worked so well in these applications is that we are able simulate millions of interactions with the environment in a relatively short period of time, allowing the agent to experience a large number of different situations in the environment and learn the consequences of its actions. In many real world applications, however, where the agent interacts with the physical world, it might not be easy to generate such a large number of interactions. The time and cost associated with training such systems could render RL an unfeasible approach for training in large scale. As a concrete example, consider training a large number of humanoid robots (agents) to move quickly, as in the Robocup competition [ Farchy et al., 2013 ] . Although the agents have similar dynamics, subtle variations mean that a single policy shared across all agents would not be an effective solution.
arXiv.org Artificial Intelligence
Jan-6-2020
- Country:
- North America > United States
- Massachusetts
- Hampshire County > Amherst (0.14)
- Plymouth County > Norwell (0.04)
- Massachusetts
- Oceania > Australia
- New South Wales > Sydney (0.04)
- North America > United States
- Genre:
- Research Report (0.64)
- Industry:
- Leisure & Entertainment > Games (0.67)
- Technology: