EgoSurgery-HTS: A Dataset for Egocentric Hand-Tool Segmentation in Open Surgery Videos
Darjana, Nathan, Fujii, Ryo, Saito, Hideo, Kajita, Hiroki
–arXiv.org Artificial Intelligence
Egocentric open-surgery videos capture rich, fine-grained details essential for accurately modeling surgical procedures and human behavior in the operating room. A detailed, pixel-level understanding of hands and surgical tools is crucial for interpreting a surgeon's actions and intentions. We introduce EgoSurgery-HTS, a new dataset with pixel-wise annotations and a benchmark suite for segmenting surgical tools, hands, and interacting tools in egocentric open-surgery videos. Specifically, we provide a labeled dataset for (1) tool instance segmentation of 14 distinct surgical tools, (2) hand instance segmentation, and (3) hand-tool segmentation to label hands and the tools they manipulate. Using EgoSurgery-HTS, we conduct extensive evaluations of state-of-the-art segmentation methods and demonstrate significant improvements in the accuracy of hand and hand-tool segmentation in egocentric open-surgery videos compared to existing datasets. The dataset will be released at https://github.com/Fujiry0/EgoSurgery.
arXiv.org Artificial Intelligence
Mar-24-2025
- Country:
- Asia > Japan
- Honshū > Kantō
- Kanagawa Prefecture > Yokohama (0.04)
- Tokyo Metropolis Prefecture > Tokyo (0.28)
- Honshū > Kantō
- Europe > Netherlands
- North Holland > Amsterdam (0.04)
- Asia > Japan
- Genre:
- Research Report (0.64)
- Industry:
- Health & Medicine > Surgery (1.00)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning (1.00)
- Vision (1.00)
- Information Technology > Artificial Intelligence