Quantifying Generalisation in Imitation Learning
Gavenski, Nathan, Rodrigues, Odinaldo
–arXiv.org Artificial Intelligence
Imitation learning benchmarks often lack sufficient variation between training and evaluation, limiting meaningful generalisation assessment. We introduce Labyrinth, a benchmarking environment designed to test generalisation with precise control over structure, start and goal positions, and task complexity. It enables verifiably distinct training, evaluation, and test settings. Labyrinth provides a discrete, fully observable state space and known optimal actions, supporting interpretability and fine-grained evaluation. Its flexible setup allows targeted testing of generalisation factors and includes variants like partial observability, key-and-door tasks, and ice-floor hazards. By enabling controlled, reproducible experiments, Labyrinth advances the evaluation of generalisation in imitation learning and provides a valuable tool for developing more robust agents.
arXiv.org Artificial Intelligence
Sep-30-2025
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe > United Kingdom
- England > Cambridgeshire > Cambridge (0.04)
- Asia > Middle East
- Genre:
- Research Report
- Experimental Study (1.00)
- New Finding (0.93)
- Research Report
- Industry:
- Education (0.93)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning
- Neural Networks (0.93)
- Reinforcement Learning (1.00)
- Representation & Reasoning > Agents (1.00)
- Robots (1.00)
- Machine Learning
- Information Technology > Artificial Intelligence