GRIM: Task-Oriented Grasping with Conditioning on Generative Examples
Shailesh, null, Raj, Alok, Kumar, Nayan, Shukla, Priya, Melnik, Andrew, Beetz, Michael, Nandi, Gora Chand
–arXiv.org Artificial Intelligence
Task-Oriented Grasping (TOG) requires robots to select grasps that are functionally appropriate for a specified task - a challenge that demands an understanding of task semantics, object affordances, and functional constraints. We present GRIM (Grasp Re-alignment via Iterative Matching), a training-free framework that addresses these challenges by leveraging Video Generation Models (VGMs) together with a retrieve-align-transfer pipeline. Beyond leveraging VGMs, GRIM can construct a memory of object-task exemplars sourced from web images, human demonstrations, or generative models. The retrieved task-oriented grasp is then transferred and refined by evaluating it against a set of geometrically stable candidate grasps to ensure both functional suitability and physical feasibility. GRIM demonstrates strong generalization and achieves state-of-the-art performance on standard TOG benchmarks. Project website: https://grim-tog.github.io
arXiv.org Artificial Intelligence
Nov-18-2025
- Genre:
- Research Report (0.50)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning (1.00)
- Natural Language > Generation (0.34)
- Representation & Reasoning > Object-Oriented Architecture (0.48)
- Robots (1.00)
- Information Technology > Artificial Intelligence