Generative Adversarial Imitation Learning
–Neural Information Processing Systems
Consider learning a policy from example expert behavior, without interaction with the expert or access to a reinforcement signal. One approach is to recover the expert's cost function with inverse reinforcement learning, then extract a policy from that cost function with reinforcement learning. This approach is indirect and can be slow. We propose a new general framework for directly extracting a policy from data as if it were obtained by reinforcement learning following inverse reinforcement learning. We show that a certain instantiation of our framework draws an analogy between imitation learning and generative adversarial networks, from which we derive a model-free imitation learning algorithm that obtains significant performance gains over existing model-free methods in imitating complex behaviors in large, high-dimensional environments.
Neural Information Processing Systems
Mar-12-2024, 17:44:43 GMT
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe
- Spain > Catalonia
- Barcelona Province > Barcelona (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Spain > Catalonia
- North America > United States
- California > Santa Clara County > Palo Alto (0.04)
- Asia > Middle East
- Technology: