When Maximum Entropy Misleads Policy Optimization
Zhang, Ruipeng, Chang, Ya-Chien, Gao, Sicun
–arXiv.org Artificial Intelligence
The Maximum Entropy Reinforcement Learning (MaxEnt RL) framework is a leading approach for achieving efficient learning and robust performance across many RL tasks. However, MaxEnt methods have also been shown to struggle with performance-critical control problems in practice, where non-MaxEnt algorithms can successfully learn. In this work, we analyze how the trade-off between robustness and optimality affects the performance of MaxEnt algorithms in complex control tasks: while entropy maximization enhances exploration and robustness, it can also mislead policy optimization, leading to failure in tasks that require precise, low-entropy policies. Through experiments on a variety of control problems, we concretely demonstrate this misleading effect. Our analysis leads to better understanding of how to balance reward design and entropy maximization in challenging control problems.
arXiv.org Artificial Intelligence
Jun-13-2025
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe > Sweden
- North America
- Canada (0.04)
- United States
- California > San Diego County
- San Diego (0.04)
- Illinois > Cook County
- Chicago (0.04)
- California > San Diego County
- Asia > Middle East
- Genre:
- Research Report (0.50)
- Industry:
- Information Technology (0.46)
- Technology: