Online Observer-Based Inverse Reinforcement Learning
Self, Ryan, Coleman, Kevin, Bai, He, Kamalapurkar, Rushikesh
–arXiv.org Artificial Intelligence
In this paper, a novel approach to the output-feedback inverse reinforcement learning (IRL) problem is developed by casting the IRL problem, for linear systems with quadratic cost functions, as a state estimation problem. Two observer-based techniques for IRL are developed, including a novel observer method that re-uses previous state estimates via history stacks. Theoretical guarantees for convergence and robustness are established under appropriate excitation conditions. Simulations demonstrate the performance of the developed observers and filters under noisy and noise-free measurements.
arXiv.org Artificial Intelligence
Jul-17-2023
- Country:
- North America > United States
- Massachusetts (0.14)
- Oklahoma (0.14)
- North America > United States
- Genre:
- Research Report > Promising Solution (0.34)
- Technology: