Goto

Collaborating Authors

 surgeon


Remote robot surgery removes cancer 1,500 miles away

FOX News

A London doctor controlled robotic surgical arms in Gibraltar to perform prostate cancer telesurgery in near real time from 1,500 miles away.


Translating music into light and motion with robots

Robohub

A system developed by researchers at the University of Waterloo lets people collaborate with groups of robots to create works of art inspired by music. The new technology features multiple wheeled robots about the size of soccer balls that trail coloured light as they move within a fixed area on the floor in response to key features of music including tempo and chord progression. A camera records the co-ordinated light trails as they snake within that area, which serves as the canvas for the creation of a "painting," or visual representation of the emotional content of a particular piece of music. "Basically, we programmed a swarm of robots to paint based on musical input," said Dr Gennaro Notomista, a professor of electrical and computer engineering at Waterloo. "The result is a cohesive system that not only processes musical input, but also co-ordinates multiple painting robots to create adaptive, expressive art that reflects the emotional essence of the music being played."


Restoring surgeons' sense of touch with robotic fingertips

Robohub

Modern surgery has gone from long incisions to tiny cuts guided by robots and AI. In the process, however, surgeons have lost something vital: the chance to feel inside the body directly. Without palpation, it becomes harder to detect tissue abnormalities during an operation. A group of surgeons and engineers across Europe is now trying to bring back this vital aspect of surgery. Working within an EU-funded research collaboration called PALPABLE, they are developing a soft robotic "fingertip" that can sense how firm or soft tissue is during minimally invasive and robotic surgery.


Disturbance-Free Surgical Video Generation from Multi-Camera Shadowless Lamps for Open Surgery

Kato, Yuna, Mori, Shohei, Saito, Hideo, Takatsume, Yoshifumi, Kajita, Hiroki, Isogawa, Mariko

arXiv.org Artificial Intelligence

Video recordings of open surgeries are greatly required for education and research purposes. However, capturing unobstructed videos is challenging since surgeons frequently block the camera field of view. To avoid occlusion, the positions and angles of the camera must be frequently adjusted, which is highly labor-intensive. Prior work has addressed this issue by installing multiple cameras on a shadowless lamp and arranging them to fully surround the surgical area. This setup increases the chances of some cameras capturing an unobstructed view. However, manual image alignment is needed in post-processing since camera configurations change every time surgeons move the lamp for optimal lighting. This paper aims to fully automate this alignment task. The proposed method identifies frames in which the lighting system moves, realigns them, and selects the camera with the least occlusion to generate a video that consistently presents the surgical field from a fixed perspective. A user study involving surgeons demonstrated that videos generated by our method were superior to those produced by conventional methods in terms of the ease of confirming the surgical area and the comfort during video viewing. Additionally, our approach showed improvements in video quality over existing techniques. Furthermore, we implemented several synthesis options for the proposed view-synthesis method and conducted a user study to assess surgeons' preferences for each option.


Advancing Minimally Invasive Precision Surgery in Open Cavities with Robotic Flexible Endoscopy

Mattille, Michelle, Mesot, Alexandre, Weisskopf, Miriam, Ochsenbein-Kölble, Nicole, Moehrlen, Ueli, Nelson, Bradley J., Boehler, Quentin

arXiv.org Artificial Intelligence

Flexible robots hold great promise for enhancing minimally invasive surgery (MIS) by providing superior dexterity, precise control, and safe tissue interaction. Yet, translating these advantages into endoscopic interventions within open cavities remains challenging. The lack of anatomical constraints and the inherent flexibility of such devices complicate their control, while the limited field of view of endoscopes restricts situational awareness. We present a robotic platform designed to overcome these challenges and demonstrate its potential in fetoscopic laser coagulation, a complex MIS procedure typically performed only by highly experienced surgeons. Our system combines a magnetically actuated flexible endoscope with teleoperated and semi-autonomous navigation capabilities for performing targeted laser ablations. To enhance surgical awareness, the platform reconstructs real-time mosaics of the endoscopic scene, providing an extended and continuous visual context. The ability of this system to address the key limitations of MIS in open spaces is validated in vivo in an ovine model.


Surgeons from Scotland and US achieve world-first stroke surgery using robot

BBC News

Doctors from Scotland and the US have completed what is thought to be a world-first stroke procedure using a robot. Prof Iris Grunwald, of the University of Dundee, performed the remote thrombectomy - the removal of blood clots after a stroke - on a human cadaver that had been donated to medical science. The professor was at Ninewells Hospital in Dundee, while the body she was operating on while using the machine was across the city at the university. Hours later, Ricardo Hanel - a neurosurgeon in Florida - used the technology to carry out the first transatlantic surgery from his Jacksonville base on a human body in Dundee over 4,000 miles (6,400km) away. The team has called it a potential game changer if it becomes approved for use on patients.


Who Sees the Risk? Stakeholder Conflicts and Explanatory Policies in LLM-based Risk Assessment

Yadav, Srishti, Gajcin, Jasmina, Miehling, Erik, Daly, Elizabeth

arXiv.org Artificial Intelligence

Understanding how different stakeholders perceive risks in AI systems is essential for their responsible deployment. This paper presents a framework for stakeholder-grounded risk assessment by using LLMs, acting as judges to predict and explain risks. Using the Risk Atlas Nexus and GloVE explanation method, our framework generates stakeholder-specific, interpretable policies that shows how different stakeholders agree or disagree about the same risks. We demonstrate our method using three real-world AI use cases of medical AI, autonomous vehicles, and fraud detection domain. We further propose an interactive visualization that reveals how and why conflicts emerge across stakeholder perspectives, enhancing transparency in conflict reasoning. Our results show that stakeholder perspectives significantly influence risk perception and conflict patterns. Our work emphasizes the importance of these stakeholder-aware explanations needed to make LLM-based evaluations more transparent, interpretable, and aligned with human-centered AI governance goals.


Kinematic and Ergonomic Design of a Robotic Arm for Precision Laparoscopic Surgery

Hao, Tian, Lu, Tong, Chan, Che

arXiv.org Artificial Intelligence

Robotic assistance in minimally invasive surgery can greatly enhance surgical precision and reduce surgeon fatigue. This paper presents a focused investigation on the kinematic and ergonomic design principles for a laparoscopic surgical robotic arm aimed at high-precision tasks. We propose a 7-degree-of-freedom (7-DOF) robotic arm system that incorporates a remote center of motion (RCM) at the instrument insertion point and ergonomic considerations to improve surgeon interaction. The design is implemented on a general-purpose robotic platform, and a series of simulated surgical tasks were performed to evaluate targeting accuracy, task efficiency, and surgeon comfort compared to conventional manual laparoscopy. Experimental results demonstrate that the optimized robotic design achieves significantly improved targeting accuracy (error reduced by over 50%) and shorter task completion times, while substantially lowering operator muscle strain and discomfort. These findings validate the importance of kinematic optimization (such as added articulations and tremor filtering) and human-centered ergonomic design in enhancing the performance of robot-assisted surgery. The insights from this work can guide the development of next-generation surgical robots that improve surgical outcomes and ergonomics for the operating team.


Control Modes of Teleoperated Surgical Robotic System's Tools in Ophthalmic Surgery

Wang, Haoran, Foroutani, Yasamin, Nepo, Matthew, Rodriguez, Mercedes, Ma, Ji, Hubschman, Jean-Pierre, Tsao, Tsu-Chin, Rosen, Jacob

arXiv.org Artificial Intelligence

Abstract--The introduction of a teleoperated surgical robotic system designed for minimally invasive procedures enables the emulation of two distinct control modes through a dedicated input device of the surgical console: (1) Inside Control Mode, which emulates tool manipulation near the distal end (i.e., as if the surgeon was holding the tip of the instrument inside the patient's body), and (2) Outside Control Mode, which emulates manipulation near the proximal end (i.e., as if the surgeon was holding the tool externally). The overarching aim of this reported research is to study and compare the surgeon's performance utilizing these two control modes of operation along with various scaling factors in a simulated vitreoretinal surgical setting. The console of Intraocular Robotic Interventional Surgical System (IRISS) was utilized but the surgical robot itself and the human eye anatomy was simulated by a virtual environment (VR) projected microscope view of an intraocular setup to a VR headset. Five experienced vitreoretinal surgeons and five subjects with no surgical experience used the system to perform fundamental tool/tissue tasks common to vitreoretinal surgery including: (1) touch and reset; (2) grasp and drop; (3) inject; (4) circular tracking. The results indicate that Inside Control outperforms Outside Control across multiple tasks and performance metrics. Higher scaling factors (20 and 30) generally provided better performance, particularly for reducing trajectory errors and tissue damage. This improvement suggests that larger scaling factors enable more precise control, making them the preferred option for fine manipulation tasks. However, task completion time was not consistently reduced across all conditions, indicating that surgeons may need to balance speed and accuracy/precision based on specific surgical requirements. By optimizing control dynamics and user interface, robotic teleoperation has the potential to reduce complications, enhance surgical dexterity, and expand the accessibility of high-precision procedures to a broader range of practitioners. In Minimally Invasive Surgery (MIS), surgical instruments are introduced into the body through small ports established at the skin surface or, in the case of ophthalmic procedures, through specific ocular tissues such as the sclera, cornea, or conjunctiva. Unlike open surgery, where the surgeon may manipulate the tool from any position along its shaft--including proximally or distally--MIS confines the surgeon's interaction to the proximal end of the tool, which remains external to the patient's body, while the distal end performs the intervention through the fixed port.


Cataract-LMM: Large-Scale, Multi-Source, Multi-Task Benchmark for Deep Learning in Surgical Video Analysis

Ahmadi, Mohammad Javad, Gandomi, Iman, Abdi, Parisa, Mohammadi, Seyed-Farzad, Taslimi, Amirhossein, Khodaparast, Mehdi, Hashemi, Hassan, Tavakoli, Mahdi, Taghirad, Hamid D.

arXiv.org Artificial Intelligence

The persistent gap between the growing global surgical demand and the trained surgical workforce [1] highlights the need to develop scalable solutions that can enhance training paradigms and optimize workflow management [2]. Computer-assisted surgery (CAS) systems are one approach to address this challenge, with applications in preoperative planning [3], intraoperative guidance [4], and standardized postoperative assessment [5, 6]. The development and validation of these advanced CAS capabilities fundamentally depend on access to large-scale, deeply annotated surgical video datasets that capture procedural phases, instrument-tissue interactions, and technical skill cues [7, 8]. Phacoemulsification cataract surgery is the most common ophthalmic procedure worldwide and the primary intervention for avoidable blindness [9, 10]. This makes it a critical domain for developing data-driven CAS with potential applications in clinical workflows and training [11, 12]. Publicly available datasets for developing CAS in cataract surgery, such as Cataract-1K [13] and CaDIS [14], are limited by their single-center origin and limited annotation scopes [15]. The absence of a multi-source dataset with comprehensive and multi-layered annotations, including objective skill assessments, has limited the development of generalizable multi-task deep learning models [11]. To address this gap, we present the Cataract-LMM (Large-scale, Multi-source, Multi-task) Dataset, a dataset of 3,000 phacoemulsification procedures recorded at two distinct clinical centers (Farabi and Noor Eye Hospitals, Tehran, Iran) between December 2021 and March 2025. The dataset is enriched with four complementary layers of annotations on subsets of the data: 1. Temporal Phase Labels (Phase): Frame-wise annotations for 13 surgical phases across 150 videos to support automated workflow recognition.