Convergence and stability of Q-learning in Hierarchical Reinforcement Learning
Manenti, Massimiliano, Iannelli, Andrea
–arXiv.org Artificial Intelligence
Decision-making architectures have played a central role for decades [1] both in engineering and other domains, e.g., guidance, navigation and control of Apollo missions [2], chemical plants [3], smart grids [4], unmanned aerial vehicles [5], recommender systems [6], and algorithms [7]. Moreover, architectures are ubiquitous in nature, e.g., diversity in the nervous system enables humans to have fast and accurate sensorimotor control [8]. Reinforcement Learning (RL) is a framework in which an agent learns to make sequential decisions through interaction with an environment in order to maximize cumulative reward [9]. Decision-making architectures have also been proposed and studied in RL. Hierarchical Reinforcement Learning (HRL) is a subfield of RL that deals with hierarchical structures for decision-making agents. Prospective advantages include improved long-term credit assignment, continual learning, interpretability, and the integration of preexisting policies [10], [11].
arXiv.org Artificial Intelligence
Nov-24-2025
- Country:
- Europe
- Germany > Baden-Württemberg
- Stuttgart Region > Stuttgart (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Germany > Baden-Württemberg
- North America > United States
- Massachusetts > Middlesex County > Cambridge (0.04)
- Europe
- Genre:
- Research Report (1.00)
- Industry:
- Energy (0.34)
- Government > Regional Government
- Materials > Chemicals (0.34)
- Technology: