LLM-based ambiguity detection in natural language instructions for collaborative surgical robots
Davila, Ana, Colan, Jacinto, Hasegawa, Yasuhisa
–arXiv.org Artificial Intelligence
Ambiguity in natural language instructions poses significant risks in safety-critical human-robot interaction, particularly in domains such as surgery. To address this, we propose a framework that uses Large Language Models (LLMs) for ambiguity detection specifically designed for collaborative surgical scenarios. Our method employs an ensemble of LLM evaluators, each configured with distinct prompting techniques to identify linguistic, contextual, procedural, and critical ambiguities. A chain-of-thought evaluator is included to systematically analyze instruction structure for potential issues. Individual evaluator assessments are synthesized through conformal prediction, which yields non-conformity scores based on comparison to a labeled calibration dataset. Evaluating Llama 3.2 11B and Gemma 3 12B, we observed classification accuracy exceeding 60% in differentiating ambiguous from unambiguous surgical instructions. Our approach improves the safety and reliability of human-robot collaboration in surgery by offering a mechanism to identify potentially ambiguous instructions before robot action.
arXiv.org Artificial Intelligence
Jul-16-2025
- Country:
- Asia > Japan (0.05)
- North America > United States (0.04)
- Genre:
- Research Report (1.00)
- Industry:
- Health & Medicine > Surgery (1.00)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning
- Neural Networks > Deep Learning (0.35)
- Performance Analysis > Accuracy (0.49)
- Natural Language > Large Language Model (1.00)
- Robots (1.00)
- Machine Learning
- Information Technology > Artificial Intelligence