Distributional Actor-Critic Ensemble for Uncertainty-Aware Continuous Control
Kanazawa, Takuya, Wang, Haiyan, Gupta, Chetan
–arXiv.org Artificial Intelligence
Uncertainty quantification is one of the central challenges for machine learning in real-world applications. In reinforcement learning, an agent confronts two kinds of uncertainty, called epistemic uncertainty and aleatoric uncertainty. Disentangling and evaluating these uncertainties simultaneously stands a chance of improving the agent's final performance, accelerating training, and facilitating quality assurance after deployment. In this work, we propose an uncertainty-aware reinforcement learning algorithm for continuous control tasks that extends the Deep Deterministic Policy Gradient algorithm (DDPG). It exploits epistemic uncertainty to accelerate exploration and aleatoric uncertainty to learn a risk-sensitive policy. We conduct numerical experiments showing that our variant of DDPG outperforms vanilla DDPG without uncertainty estimation in benchmark tasks on robotic control and power-grid optimization.
arXiv.org Artificial Intelligence
Jul-27-2022
- Country:
- Asia (0.67)
- Europe (1.00)
- North America
- Canada (0.68)
- United States > California (0.68)
- Genre:
- Overview (0.46)
- Research Report (0.64)
- Industry:
- Energy
- Power Industry (1.00)
- Renewable > Solar (0.46)
- Energy
- Technology: