Tracking Everything in Robotic-Assisted Surgery

Bohan Zhan1, Wang Zhao2, Yi Fang3, Bo Du4, Francisco Vasconcelos5, Danail Stoyanov5, Daniel S. Elson1, Baoru Huang1,5,6*,
1Imperial College London, 2Tsinghua University, 3NYU Abu Dhabi,
4Wuhan University, 5University College London, 6University of Liverpool
Code arXiv

Abstract

Accurate tracking of tissues and instruments in videos is crucial for Robotic-Assisted Minimally Invasive Surgery (RAMIS), as it enables the robot to comprehend the surgical scene with precise locations and interactions of tissues and tools. Traditional keypoint-based sparse tracking is limited by featured points, while flow-based dense two-view matching suffers from long-term drifts. Recently, the Tracking Any Point (TAP) algorithm was proposed to overcome these limitations and achieve dense accurate long-term tracking. However, its efficacy in surgical scenarios remains untested, largely due to the lack of a comprehensive surgical tracking dataset for evaluation. To address this gap, we introduce a new annotated surgical tracking dataset for benchmarking tracking methods for surgical scenarios, comprising real-world surgical videos with complex tissue and instrument motions. We extensively evaluate state-of-the-art (SOTA) TAP-based algorithms on this dataset and reveal their limitations in challenging surgical scenarios, including fast instrument motion, severe occlusions, and motion blur, etc. Furthermore, we propose a new tracking method, namely SurgMotion, to solve the challenges and further improve the tracking performance. Our proposed method outperforms most TAP-based algorithms in surgical instruments tracking, and especially demonstrates significant improvements over baselines in challenging medical videos.

Dataset

Method Overview

Method Overview Diagram

Building upon OmniMotion, we proposed three key constraints: the tool mask constraint, the as-rigid-as-possible (ARAP) constraint, and sparse feature matching guidance.

Result

Comparison

BibTeX


@article{zhan2024tracking,
  title={Tracking Everything in Robotic-Assisted Surgery},
  author={Zhan, Bohan and Zhao, Wang and Fang, Yi and Du, Bo and Vasconcelos, Francisco and Stoyanov, Danail and Elson, Daniel S and Huang, Baoru},
  journal={arXiv preprint arXiv:2409.19821},
  year={2024}
}