Instrument-Splatting: Controllable Photorealistic Reconstruction of Surgical Instruments Using Gaussian Splatting

📅 2025-03-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address low fidelity and poor controllability in 3D reconstruction of surgical instruments from monocular surgical videos, this paper proposes the first instrument-level controllable 3D reconstruction framework tailored for Real2Sim applications. Methodologically, we introduce a novel geometry pretraining strategy that binds Gaussian point clouds with part-wise meshes; integrate forward kinematics modeling to enable joint-level controllable deformation; and design a semantic-embedding Gaussian rendering-contrastive pose tracking scheme that jointly optimizes real-time pose and joint states. Evaluated on six real surgical videos—including both public and in-house datasets—our method achieves photorealistic rendering and millimeter-level geometric accuracy (mean error <1.2 mm), significantly outperforming NeRF-based and generic point-cloud approaches. This work establishes a high-fidelity, physics-aware, and actuation-ready 3D foundation for surgical AI simulation and training.

Technology Category

Application Category

📝 Abstract
Real2Sim is becoming increasingly important with the rapid development of surgical artificial intelligence (AI) and autonomy. In this work, we propose a novel Real2Sim methodology, extit{Instrument-Splatting}, that leverages 3D Gaussian Splatting to provide fully controllable 3D reconstruction of surgical instruments from monocular surgical videos. To maintain both high visual fidelity and manipulability, we introduce a geometry pre-training to bind Gaussian point clouds on part mesh with accurate geometric priors and define a forward kinematics to control the Gaussians as flexible as real instruments. Afterward, to handle unposed videos, we design a novel instrument pose tracking method leveraging semantics-embedded Gaussians to robustly refine per-frame instrument poses and joint states in a render-and-compare manner, which allows our instrument Gaussian to accurately learn textures and reach photorealistic rendering. We validated our method on 2 publicly released surgical videos and 4 videos collected on ex vivo tissues and green screens. Quantitative and qualitative evaluations demonstrate the effectiveness and superiority of the proposed method.
Problem

Research questions and friction points this paper is trying to address.

Controllable 3D reconstruction of surgical instruments from monocular videos.
High visual fidelity and manipulability in surgical instrument modeling.
Robust pose tracking and photorealistic rendering in unposed surgical videos.
Innovation

Methods, ideas, or system contributions that make the work stand out.

3D Gaussian Splatting for surgical instrument reconstruction
Geometry pre-training with accurate geometric priors
Semantics-embedded Gaussians for robust pose tracking
🔎 Similar Papers
No similar papers found.
S
Shuojue Yang
National University of Singapore, Singapore, Singapore
Z
Zijian Wu
The University of British Columbia, Vancouver, Canada
M
Mingxuan Hong
National University of Singapore, Singapore, Singapore
Q
Qian Li
National University of Singapore, Singapore, Singapore
Daiyun Shen
Daiyun Shen
PhD at National University of Singapore
medical AI
S
S. Salcudean
The University of British Columbia, Vancouver, Canada
Yueming Jin
Yueming Jin
Assistant Professor, National University of Singapore
Medical Image AnalysisSurgical AI&RoboticsMultimodal Learning