Tsallis Entropy Regularization for Linearly Solvable MDP and Linear Quadratic Regulator
Hashizume, Yota, Oishi, Koshi, Kashima, Kenji
–arXiv.org Artificial Intelligence
Shannon entropy regularization is widely adopted in optimal control due to its ability to promote exploration and enhance robustness, e.g., maximum entropy reinforcement learning known as Soft Actor-Critic. In this paper, Tsallis entropy, which is a one-parameter extension of Shannon entropy, is used for the regularization of linearly solvable MDP and linear quadratic regulators. We derive the solution for these problems and demonstrate its usefulness in balancing between exploration and sparsity of the obtained control law.
arXiv.org Artificial Intelligence
Mar-4-2024
- Country:
- Asia > Japan
- Honshū > Kansai > Kyoto Prefecture > Kyoto (0.05)
- Europe > United Kingdom
- England > Oxfordshire > Oxford (0.04)
- North America > United States
- New York (0.04)
- Oceania > Australia
- South Australia > Adelaide (0.04)
- Asia > Japan
- Genre:
- Research Report (0.64)
- Technology: