Žefran, Miloš
Autonomous Dissection in Robotic Cholecystectomy
Oh, Ki-Hwan, Borgioli, Leonardo, Žefran, Miloš, Valle, Valentina, Giulianotti, Pier Cristoforo
Robotic surgery offers enhanced precision and adaptability, paving the way for automation in surgical interventions. Cholecystectomy, the gallbladder removal, is particularly well-suited for automation due to its standardized procedural steps and distinct anatomical boundaries. A key challenge in automating this procedure is dissecting with accuracy and adaptability. This paper presents a vision-based autonomous robotic dissection architecture that integrates real-time segmentation, keypoint detection, grasping and stretching the gallbladder with the left arm, and dissecting with the other. We introduce an improved segmentation dataset based on videos of robotic cholecystectomy performed by various surgeons, incorporating a new ``liver bed'' class to enhance boundary tracking after multiple rounds of dissection. Our system employs state-of-the-art segmentation models and an adaptive boundary extraction method that maintains accuracy despite tissue deformations and visual variations. Moreover, we implemented an automated grasping and pulling strategy to optimize tissue tension before dissection upon our previous work. Ex vivo evaluations on porcine livers demonstrate that our framework significantly improves dissection precision and consistency, marking a step toward fully autonomous robotic cholecystectomy.
Expanded Comprehensive Robotic Cholecystectomy Dataset (CRCD)
Oh, Ki-Hwan, Borgioli, Leonardo, Mangano, Alberto, Valle, Valentina, Di Pangrazio, Marco, Toti, Francesco, Pozza, Gioia, Ambrosini, Luciano, Ducas, Alvaro, Žefran, Miloš, Chen, Liaohai, Giulianotti, Pier Cristoforo
In recent years, the application of machine learning to minimally invasive surgery (MIS) has attracted considerable interest. Datasets are critical to the use of such techniques. This paper presents a unique dataset recorded during ex vivo pseudo-cholecystectomy procedures on pig livers using the da Vinci Research Kit (dVRK). Unlike existing datasets, it addresses a critical gap by providing comprehensive kinematic data, recordings of all pedal inputs, and offers a time-stamped record of the endoscope's movements. This expanded version also includes segmentation and keypoint annotations of images, enhancing its utility for computer vision applications. Contributed by seven surgeons with varied backgrounds and experience levels that are provided as a part of this expanded version, the dataset is an important new resource for surgical robotics research. It enables the development of advanced methods for evaluating surgeon skills, tools for providing better context awareness, and automation of surgical tasks. Our work overcomes the limitations of incomplete recordings and imprecise kinematic data found in other datasets. To demonstrate the potential of the dataset for advancing automation in surgical robotics, we introduce two models that predict clutch usage and camera activation, a 3D scene reconstruction example, and the results from our keypoint and segmentation models.
Group-Control Motion Planning Framework for Microrobot Swarms in a Global Field
Li, Siyu, Shervedani, Afagh Mehri, Žefran, Miloš, Paprotny, Igor
This paper investigates how group-control can be effectively used for motion planning for microrobot swarms in a global field. We prove that Small-Time Local Controllability (STLC) in robot positions is achievable through group-control, with the minimum number of groups required for STLC being $\log_2(n + 2) + 1$ for $n$ robots. We then discuss the complexity trade-offs between control and motion planning. We show how motion planning can be simplified if appropriate primitives can be achieved through more complex control actions. We identify motion planning problems that balance the number of robot groups and motion primitives with planning complexity. Various instantiations of these motion planning problems are explored, with simulations to demonstrate the effectiveness of group-control.
The RoboHelper Project: From Multimodal Corpus to Embodiment on a Robot
Eugenio, Barbara Di (University of Illinois Chicago) | Žefran, Miloš (University of Illinois at Chicago)
In this position paper, we describe the RoboHelper project, its findings and our vision for its future. The long-term goal of RoboHelper is to develop assistive robots for the elderly. The main thesis of our work is that such robots must crucially be able to participate in multimodal dialogues. Contributions of our work to date include the ELDERLY-AT-HOME corpus that we collected and annotated. It consists of 20 task-oriented human-human dialogues between a helper and an elderly person in a fully functional apartment. The unique feature of the corpus is that in addition to video and audio, it includes recordings of physical interaction. Based on this data, we have demonstrated the crucial role that Haptic-Ostensive (H-O) actions play in interpreting language and uncovering a person's intentions. H-O actions manipulate objects, but they also often perform a referring function. Our models were derived on the basis of manually annotated categories. Additional experiments show that we can identify H-O actions using the physical interaction data measured through an unobtrusive sensory glove developed as part of the project. In future work, we will derive models for the robot to decide what to do next (as opposed to interpreting what the interlocutor did); explore other types of physical interactions; and refine preliminary implementations of our models on the Nao robotic platform.