Training-free Task-oriented Grasp Generation
Wang, Jiaming, Chen, Jizhuo, Liu, Diwen
–arXiv.org Artificial Intelligence
This paper presents a training-free pipeline for task-oriented grasp generation that combines pre-trained grasp generation models with vision-language models (VLMs). Unlike traditional approaches that focus solely on stable grasps, our method incorporates task-specific requirements by leveraging the semantic reasoning capabilities of VLMs. We evaluate five querying strategies, each utilizing different visual representations of candidate grasps, and demonstrate significant improvements over a baseline method in both grasp success and task compliance rates, with absolute gains of up to 36.9% in overall success rate. Our results underline the potential of VLMs to enhance task-oriented manipulation, providing insights for future research in robotic grasping and human-robot interaction.
arXiv.org Artificial Intelligence
Feb-7-2025
- Genre:
- Research Report > New Finding (0.34)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning
- Neural Networks (0.46)
- Statistical Learning (0.47)
- Natural Language (1.00)
- Robots (1.00)
- Machine Learning
- Information Technology > Artificial Intelligence