Learning to Cooperate with Humans using Generative Agents
–Neural Information Processing Systems
Training agents that can coordinate zero-shot with humans is a key mission in multi-agent reinforcement learning (MARL). Current algorithms focus on training simulated human partner policies which are then used to train a Cooperator agent. The simulated human is produced either through behavior cloning over a dataset of human cooperation behavior, or by using MARL to create a population of simulated agents. However, these approaches often struggle to produce a Cooperator that can coordinate well with real humans, since the simulated humans fail to cover the diverse strategies and styles employed by people in the real world. We show learning a generative model of human partners can effectively address this issue.
Neural Information Processing Systems
May-29-2025, 22:43:51 GMT
- Country:
- North America > Canada (0.14)
- Genre:
- Questionnaire & Opinion Survey (0.93)
- Research Report
- Experimental Study (1.00)
- New Finding (0.93)
- Industry:
- Leisure & Entertainment > Games (1.00)
- Technology: