Robust Autonomy Emerges from Self-Play
Cusumano-Towner, Marco, Hafner, David, Hertzberg, Alex, Huval, Brody, Petrenko, Aleksei, Vinitsky, Eugene, Wijmans, Erik, Killian, Taylor, Bowers, Stuart, Sener, Ozan, Krähenbühl, Philipp, Koltun, Vladlen
–arXiv.org Artificial Intelligence
Self-play has powered breakthroughs in two-player and multi-player games. Here we show that self-play is a surprisingly effective strategy in another domain. We show that robust and naturalistic driving emerges entirely from self-play in simulation at unprecedented scale -- 1.6~billion~km of driving. This is enabled by Gigaflow, a batched simulator that can synthesize and train on 42 years of subjective driving experience per hour on a single 8-GPU node. The resulting policy achieves state-of-the-art performance on three independent autonomous driving benchmarks. The policy outperforms the prior state of the art when tested on recorded real-world scenarios, amidst human drivers, without ever seeing human data during training. The policy is realistic when assessed against human references and achieves unprecedented robustness, averaging 17.5 years of continuous driving between incidents in simulation.
arXiv.org Artificial Intelligence
Feb-5-2025
- Country:
- Asia
- Middle East > Jordan (0.04)
- Singapore (0.04)
- Europe > United Kingdom
- England > Greater London > London (0.04)
- North America > United States
- California
- San Diego County > San Diego (0.04)
- San Francisco County > San Francisco (0.04)
- Michigan (0.04)
- Nevada > Clark County
- Las Vegas (0.04)
- California
- Asia
- Genre:
- Research Report (0.82)
- Industry:
- Automobiles & Trucks (1.00)
- Leisure & Entertainment > Games (1.00)
- Transportation
- Ground > Road (1.00)
- Infrastructure & Services (1.00)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning
- Neural Networks (0.93)
- Reinforcement Learning (0.68)
- Representation & Reasoning > Agents (1.00)
- Robots (1.00)
- Machine Learning
- Information Technology > Artificial Intelligence