Multimodal Image-Guided Robotic Surgery (MIGRS): A Novel Unified Framework for Real-Time Navigation and Visualization
Ehsan Nasiri
Ehsan Nasiri
This MIGRS, is a novel unified multimodal image-guided robotic surgery framework for real-time visualization and navigation. It integrates stereo 3D reconstruction (point clouds), AR video overlay, TilePro HUD on the da Vinci console, navigation/visualization, and ultrasound, with modules for stereo calibration, tracker registration, and fiducial-based rigid registration with live RMSE feedback. High-throughput video I/O and rendering run in C++, while a Python GUI interfaces via a zero-copy novel binding wrapper; stereo streaming and optical tracking operate in real time. Across 18 experiments on da Vinci S, Xi, and SP systems, MIGRS auto-adapts to platform-specific endoscopes and video formats. Stereo inference averaged ~0.20 s per pair. Over clinically relevant working distances, mean 3D reprojection error was ~0.6 mm, with best 0.1 mm RMSE at 100 mm on Xi. Results show responsive, accurate, extensible multimodal guidance for research and translational workflows.
A U.S. provisional patent application covering aspects of this work has been filed (U.S. provisional application No. 63/920,165).