Reward Dimension Reduction for Scalable Multi-Objective Reinforcement Learning
Park, Giseung, Sung, Youngchul
–arXiv.org Artificial Intelligence
In this paper, we introduce a simple yet effective reward dimension reduction method to tackle the scalability challenges of multi-objective reinforcement learning algorithms. While most existing approaches focus on optimizing two to four objectives, their abilities to scale to environments with more objectives remain uncertain. Our method uses a dimension reduction approach to enhance learning efficiency and policy performance in multi-objective settings. While most traditional dimension reduction methods are designed for static datasets, our approach is tailored for online learning and preserves Pareto-optimality after transformation. We propose a new training and evaluation framework for reward dimension reduction in multi-objective reinforcement learning and demonstrate the superiority of our method in environments including one with sixteen objectives, significantly outperforming existing online dimension reduction methods.
arXiv.org Artificial Intelligence
Feb-28-2025
- Country:
- Africa > Rwanda
- Asia > South Korea
- Europe
- North America
- Canada
- Alberta > Census Division No. 15
- Improvement District No. 9 > Banff (0.04)
- British Columbia > Vancouver (0.04)
- Alberta > Census Division No. 15
- Mexico > Guanajuato (0.04)
- United States
- California
- Los Angeles County > Long Beach (0.04)
- San Diego County > San Diego (0.04)
- Nevada (0.04)
- California
- Canada
- Genre:
- Research Report (1.00)
- Industry:
- Education > Educational Setting > Online (0.48)
- Technology: