No More Marching: Learning Humanoid Locomotion for Short-Range SE(2) Targets
Dugar, Pranay, Gadde, Mohitvishnu S., Siekmann, Jonah, Godse, Yesh, Shrestha, Aayam, Fern, Alan
–arXiv.org Artificial Intelligence
Humanoids operating in real-world workspaces must frequently execute task-driven, short-range movements to SE(2) target poses. To be practical, these transitions must be fast, robust, and energy efficient. While learning-based locomotion has made significant progress, most existing methods optimize for velocity-tracking rather than direct pose reaching, resulting in inefficient, marching-style behavior when applied to short-range tasks. In this work, we develop a reinforcement learning approach that directly optimizes humanoid locomotion for SE(2) targets. Central to this approach is a new constellation-based reward function that encourages natural and efficient target-oriented movement. To evaluate performance, we introduce a benchmarking framework that measures energy consumption, time-to-target, and footstep count on a distribution of SE(2) goals. Our results show that the proposed approach consistently outperforms standard methods and enables successful transfer from simulation to hardware, highlighting the importance of targeted reward design for practical short-range humanoid locomotion.
arXiv.org Artificial Intelligence
Aug-21-2025
- Country:
- Asia > Japan
- Honshū > Chūbu > Ishikawa Prefecture > Kanazawa (0.04)
- North America > United States
- Oregon
- Benton County > Corvallis (0.04)
- Marion County > Salem (0.04)
- Oregon
- Asia > Japan
- Genre:
- Research Report > New Finding (0.86)
- Industry:
- Energy (0.35)
- Technology: