SurgSync: Time-Synchronized Multi-Modal Data Collection Framework and Dataset for Surgical Robotics

📅 2026-03-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the scarcity of high-quality, multimodal, and temporally synchronized training data that hinders the advancement of intelligent surgical robots. To this end, we present a multimodal data acquisition framework supporting both offline and online synchronization, integrating a clinical-grade stereo endoscope, lateral-view cameras, and capacitive tactile sensors, all implemented on the dVRK platform to enable high-fidelity, synchronized recording of visual, kinematic, and contact force data. We introduce a novel dual-mode synchronization mechanism and leverage techniques including stereo vision, Gaussian heatmap-based kinematic reprojection, depth estimation, and optical flow to construct a dataset comprising 214 validated multitask surgical manipulation instances performed on real tissue across varying skill levels. This dataset effectively facilitates the training of surgical skill assessment networks, and all hardware, software, and data are publicly released.

Technology Category

Application Category

📝 Abstract
Most existing robotic surgery systems adopt a human-in-the-loop paradigm, often with the surgeon directly teleoperating the robotic system. Adding intelligence to these robots would enable higher-level control, such as supervised autonomy or even full autonomy. However, artificial intelligence (AI) requires large amounts of training data, which is currently lacking. This work proposes SurgSync, a multi-modal data collection framework with offline and online synchronization to support training and real-time inference, respectively. The framework is implemented on a da Vinci Research Kit (dVRK) and introduces (1) dual-mode (online/offline-matching) synchronized recorders, (2) a modern stereo endoscope to achieve image quality on par with clinical systems, and (3) additional sensors such as a side-view camera and a novel capacitive contact sensor to provide ground truth contact data. The framework also incorporates a post-processing toolbox for tasks such as depth estimation, optical flow, and a practical kinematic reprojection method using Gaussian heatmap. User studies with participants of varying skill levels are performed with ex-vivo tissue to provide clinically realistic data, and a network for surgical skill assessment is employed to demonstrate utilization of the collected data. Through the user study experiments, we obtained a dataset of 214 validated instances across multiple canonical training tasks. All software and data are available at surgsync.github.io.
Problem

Research questions and friction points this paper is trying to address.

surgical robotics
multi-modal data
time synchronization
AI training data
autonomous surgery
Innovation

Methods, ideas, or system contributions that make the work stand out.

multi-modal synchronization
surgical robotics dataset
capacitive contact sensing
stereo endoscopy
kinematic reprojection
🔎 Similar Papers
No similar papers found.
H
Haoying Zhou
Department of Robotics Engineering, Worcester Polytechnic Institute, Worcester, MA, USA
C
Chang Liu
Laboratory for Computational Sensing and Robotics, Johns Hopkins University, Baltimore, MD, USA
Yimeng Wu
Yimeng Wu
Huawei Noah's Ark Lab
Large Language Models
J
Junlin Wu
Laboratory for Computational Sensing and Robotics, Johns Hopkins University, Baltimore, MD, USA; Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
Zijian Wu
Zijian Wu
University of British Columbia
Surgical RoboticsImage Guided SurgeryRobot-assisted Surgery
Y
Yu Chung Lee
Robotics and Control Laboratory, the University of British Columbia, Vancouver, Canada
S
Sara Martuscelli
Department of Electronics, Information and Bioengineering, Politecnico di Milano, Milano, Italy
S
Septimiu E. Salcudean
Robotics and Control Laboratory, the University of British Columbia, Vancouver, Canada
Gregory S. Fischer
Gregory S. Fischer
Professor, Worcester Polytechnic Institute (WPI), AIM Lab, PracticePoint Healthcare CPS R&D Center
Medical RoboticsComputer Assisted SurgeryMRI Compatible RoboticsImage Guided Surgery
Peter Kazanzides
Peter Kazanzides
Johns Hopkins University
roboticsmedical roboticssoftware engineering