Model-Reference Reinforcement Learning Control of Autonomous Surface Vehicles with Uncertainties
Zhang, Qingrui, Pan, Wei, Reppa, Vasso
–arXiv.org Artificial Intelligence
This paper presents a novel model-reference reinforcement learning control method for uncertain autonomous surface vehicles. The proposed control combines a conventional control method with deep reinforcement learning. With the conventional control, we can ensure the learning-based control law provides closed-loop stability for the overall system, and potentially increase the sample efficiency of the deep reinforcement learning. With the reinforcement learning, we can directly learn a control law to compensate for modeling uncertainties. In the proposed control, a nominal system is employed for the design of a baseline control law using a conventional control approach. The nominal system also defines the desired performance for uncertain autonomous vehicles to follow. In comparison with traditional deep reinforcement learning methods, our proposed learning-based control can provide stability guarantees and better sample efficiency. We demonstrate the performance of the new algorithm via extensive simulation results.
arXiv.org Artificial Intelligence
Mar-30-2020
- Country:
- Europe > Netherlands
- South Holland > Delft (0.05)
- North America
- Canada > British Columbia
- Mexico > Quintana Roo
- Cancún (0.04)
- United States
- California > Los Angeles County
- Long Beach (0.04)
- Missouri > St. Louis County
- St. Louis (0.04)
- California > Los Angeles County
- Europe > Netherlands
- Genre:
- Research Report (0.70)
- Industry:
- Energy (0.35)
- Technology: