Learning from Demonstrations using Signal Temporal Logic
Puranic, Aniruddh G., Deshmukh, Jyotirmoy V., Nikolaidis, Stefanos
–arXiv.org Artificial Intelligence
Learning-from-demonstrations is an emerging paradigm to obtain effective robot control policies for complex tasks via reinforcement learning without the need to explicitly design reward functions. However, it is susceptible to imperfections in demonstrations and also raises concerns of safety and interpretability in the learned control policies. To address these issues, we use Signal Temporal Logic to evaluate and rank the quality of demonstrations. Temporal logic-based specifications allow us to create non-Markovian rewards, and also define interesting causal dependencies between tasks such as sequential task specifications. We validate our approach through experiments on discrete-world and OpenAI Gym environments, and show that our approach outperforms the state-of-the-art Maximum Causal Entropy Inverse Reinforcement Learning.
arXiv.org Artificial Intelligence
Feb-15-2021
- Country:
- Asia
- China (0.04)
- Macao (0.04)
- Middle East
- Jordan (0.04)
- Republic of Türkiye > Aksaray Province
- Aksaray (0.04)
- Europe
- Slovenia > Upper Carniola
- Municipality of Bled > Bled (0.04)
- Sweden > Stockholm
- Stockholm (0.04)
- Slovenia > Upper Carniola
- North America > United States
- California
- Los Angeles County > Los Angeles (0.14)
- Santa Clara County > Stanford (0.04)
- Colorado > Denver County
- Denver (0.04)
- Illinois > Cook County
- Chicago (0.04)
- Massachusetts > Middlesex County
- Cambridge (0.04)
- New York > New York County
- New York City (0.04)
- Tennessee > Davidson County
- Nashville (0.04)
- California
- Oceania > Australia
- Asia
- Genre:
- Research Report (0.40)
- Technology: