Collaborative surgical instrument segmentation for monocular depth estimation in minimally invasive surgery

Xue Li, Wenxin Chen, Xingguang Duan, Xiaoyi Gu*, Changsheng Li

*此作品的通讯作者

科研成果: 期刊稿件文章同行评审

摘要

Depth estimation is essential for image-guided surgical procedures, particularly in minimally invasive environments where accurate 3D perception is critical. This paper proposes a two-stage self-supervised monocular depth estimation framework that incorporates instrument segmentation as a task-level prior to enhance spatial understanding. In the first stage, segmentation and depth estimation models are trained separately on the RIS, SCARED datasets to capture task-specific representations. In the second stage, segmentation masks predicted on the dVPN dataset are fused with RGB inputs to guide the refinement of depth prediction. The framework employs a shared encoder and multiple decoders to enable efficient feature sharing across tasks. Comprehensive experiments on the RIS, SCARED, dVPN, and SERV-CT datasets validate the effectiveness and generalizability of the proposed approach. The results demonstrate that segmentation-aware depth estimation improves geometric reasoning in challenging surgical scenes, including those with occlusions, specularities regions.

源语言英语
文章编号103765
期刊Medical Image Analysis
106
DOI
出版状态已出版 - 12月 2025
已对外发布

指纹

探究 'Collaborative surgical instrument segmentation for monocular depth estimation in minimally invasive surgery' 的科研主题。它们共同构成独一无二的指纹。

引用此