Creating Hierarchical Dispositions of Needs in an Agent
–arXiv.org Artificial Intelligence
We present a novel method for learning hierarchical abstractions that prioritize competing objectives, leading to improved global expected rewards. Our approach employs a secondary rewarding agent with multiple scalar outputs, each associated with a distinct level of abstraction. The traditional agent then learns to maximize these outputs in a hierarchical manner, conditioning each level on the maximization of the preceding level. We derive an equation that orders these scalar values and the global reward by priority, inducing a hierarchy of needs that informs goal formation. Experimental results on the Pendulum v1 environment demonstrate superior performance compared to a baseline implementation.We achieved state of the art results.
arXiv.org Artificial Intelligence
Nov-23-2024
- Country:
- Africa > Zimbabwe
- Asia > Middle East
- Jordan (0.05)
- North America > United States (0.04)
- Genre:
- Research Report
- New Finding (0.47)
- Promising Solution (0.34)
- Research Report
- Technology: