Presentation + Paper
29 March 2024 WS-SfMLearner: self-supervised monocular depth and ego-motion estimation on surgical videos with unknown camera parameters
Author Affiliations +
Abstract
Depth estimation in surgical video plays a crucial role in many image-guided surgery procedures. However, it is difficult and time consuming to create depth map ground truth datasets in surgical videos due in part to inconsistent brightness and noise in the surgical scene. Therefore, building an accurate and robust self-supervised depth and camera ego-motion estimation system is gaining more attention from the computer vision community. Although several self-supervision methods alleviate the need for ground truth depth maps and poses, they still need known camera intrinsic parameters, which are often missing or not recorded. Moreover, the camera intrinsic prediction methods in existing works depend heavily on the quality of datasets. In this work, we aimed to build a self-supervised depth and ego-motion estimation system which can predict not only accurate depth maps and camera pose, but also camera intrinsic parameters. We proposed a cost-volume-based supervision manner to give the system auxiliary supervision for camera parameters prediction. The experimental results showed that the proposed method improved the accuracy of estimated camera parameters, ego-motion, and depth estimation.
Conference Presentation
(2024) Published by SPIE. Downloading of the abstract is permitted for personal use only.
Ange Lou and Jack Noble "WS-SfMLearner: self-supervised monocular depth and ego-motion estimation on surgical videos with unknown camera parameters", Proc. SPIE 12928, Medical Imaging 2024: Image-Guided Procedures, Robotic Interventions, and Modeling, 129280L (29 March 2024); https://doi.org/10.1117/12.3008390
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Cameras

Depth maps

Video

Back to Top