deepreinforcementlearni...
- Asia > China > Beijing > Beijing (0.05)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- Asia > Middle East > Jordan (0.04)
TowardsPlayingFullMOBAGameswith DeepReinforcementLearning
As aresult, full MOBAgames without restrictions are farfrom being mastered by any existing AI system. In this paper, we propose a MOBA AIlearning paradigm that methodologically enables playing full MOBAgames withdeepreinforcementlearning.Specifically,wedevelopacombinationofnovel and existing learning techniques, including curriculum self-play learning, policy distillation, off-policy adaption, multi-head value estimation, and Monte-Carlo tree-search, intraining andplaying alargepoolofheroes,meanwhile addressing thescalabilityissueskillfully.
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- Asia > Middle East > Jordan (0.04)
- Asia > China > Sichuan Province > Chengdu (0.04)
- Asia > China > Guangdong Province > Shenzhen (0.04)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.48)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (0.47)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Search (0.34)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (0.30)
Multi-fidelity Reinforcement Learning Control for Complex Dynamical Systems
Sun, Luning, Liu, Xin-Yang, Zhao, Siyan, Grover, Aditya, Wang, Jian-Xun, Thiagarajan, Jayaraman J.
Controlling instabilities in complex dynamical systems is challenging in scientific and engineering applications. Deep reinforcement learning (DRL) has seen promising results for applications in different scientific applications. The many-query nature of control tasks requires multiple interactions with real environments of the underlying physics. However, it is usually sparse to collect from the experiments or expensive to simulate for complex dynamics. Alternatively, controlling surrogate modeling could mitigate the computational cost issue. However, a fast and accurate learning-based model by offline training makes it very hard to get accurate pointwise dynamics when the dynamics are chaotic. To bridge this gap, the current work proposes a multi-fidelity reinforcement learning (MFRL) framework that leverages differentiable hybrid models for control tasks, where a physics-based hybrid model is corrected by limited high-fidelity data. We also proposed a spectrum-based reward function for RL learning. The effect of the proposed framework is demonstrated on two complex dynamics in physics. The statistics of the MFRL control result match that computed from many-query evaluations of the high-fidelity environments and outperform other SOTA baselines.