Interpretable Concept Bottlenecks to Align Reinforcement Learning Agents Quentin Delfosse,1 Sebastian Sztwiertnia,1 Mark Rothermel 1 Wolfgang Stammer
–Neural Information Processing Systems
Goal misalignment, reward sparsity and difficult credit assignment are only a few of the many issues that make it difficult for deep reinforcement learning (RL) agents to learn optimal policies. Unfortunately, the black-box nature of deep neural networks impedes the inclusion of domain experts for inspecting the model and revising suboptimal policies. To this end, we introduce Successive Concept Bottleneck Agents (SCoBots), that integrate consecutive concept bottleneck (CB) layers. In contrast to current CB models, SCoBots do not just represent concepts as properties of individual objects, but also as relations between objects which is crucial for many RL tasks.
Neural Information Processing Systems
May-30-2025, 07:43:41 GMT
- Country:
- North America > United States > California > San Francisco County > San Francisco (0.14)
- Genre:
- Research Report
- Experimental Study (1.00)
- New Finding (0.67)
- Research Report
- Industry:
- Leisure & Entertainment > Games (0.93)
- Technology: