Decision Theory-Guided Deep Reinforcement Learning for Fast Learning
Wan, Zelin, Cho, Jin-Hee, Zhu, Mu, Anwar, Ahmed H., Kamhoua, Charles, Singh, Munindar P.
–arXiv.org Artificial Intelligence
This paper introduces a novel approach, Decision Theory-guided Deep Reinforcement Learning (DT-guided DRL), to address the inherent cold start problem in DRL. By integrating decision theory principles, DT-guided DRL enhances agents' initial performance and robustness in complex environments, enabling more efficient and reliable convergence during learning. Our investigation encompasses two primary problem contexts: the cart pole and maze navigation challenges. Experimental results demonstrate that the integration of decision theory not only facilitates effective initial guidance for DRL agents but also promotes a more structured and informed exploration strategy, particularly in environments characterized by large and intricate state spaces. The results of experiment demonstrate that DT-guided DRL can provide significantly higher rewards compared to regular DRL. Specifically, during the initial phase of training, the DT-guided DRL yields up to an 184% increase in accumulated reward. Moreover, even after reaching convergence, it maintains a superior performance, ending with up to 53% more reward than standard DRL in large maze problems. DT-guided DRL represents an advancement in mitigating a fundamental challenge of DRL by leveraging functions informed by human (designer) knowledge, setting a foundation for further research in this promising interdisciplinary domain.
arXiv.org Artificial Intelligence
Feb-8-2024
- Country:
- Europe > Austria
- Vienna (0.14)
- North America > United States
- North Carolina (0.14)
- Virginia (0.14)
- Europe > Austria
- Genre:
- Overview > Innovation (0.34)
- Research Report
- New Finding (0.48)
- Promising Solution (0.48)
- Industry:
- Automobiles & Trucks (0.47)
- Education (0.46)
- Energy > Oil & Gas
- Upstream (0.34)
- Leisure & Entertainment (0.46)
- Transportation > Ground
- Road (0.47)
- Technology: