Articulated Multi-Instrument 2-D Pose Estimation Using Fully Convolutional Networks

被引:100
作者
Du, Xiaofei [1 ]
Kurmann, Thomas [2 ]
Chang, Ping-Lin [3 ]
Allan, Maximilian [1 ]
Ourselin, Sebastien [1 ]
Sznitman, Raphael [2 ]
Kelly, John D. [4 ]
Stoyanov, Danail [1 ]
机构
[1] UCL, Ctr Med Image Comp, London WC1E 6BT, England
[2] Univ Bern, ARTORG Ctr Biomed Engn Res, CH-3012 Bern, Switzerland
[3] Umbo Comp Vis Inc, San Francisco, CA 94105 USA
[4] UCL, Div Surg & Intervent Sci, London WC1E 6BT, England
基金
英国工程与自然科学研究理事会; 英国惠康基金; 欧盟地平线“2020”;
关键词
Surgical instrument detection; articulated pose estimation; fully convolutional networks; surgical vision; SURGICAL TOOL DETECTION; 3D TRACKING;
D O I
10.1109/TMI.2017.2787672
中图分类号
TP39 [计算机的应用];
学科分类号
080201 [机械制造及其自动化];
摘要
Instrument detection, pose estimation, and tracking in surgical videos are an important vision component for computer-assisted interventions. While significant advances have been made in recent years, articulation detection is still a major challenge. In this paper, we propose a deep neural network for articulated multi-instrument 2-D pose estimation, which is trained on detailed annotations of endoscopic and microscopic data sets. Our model is formed by a fully convolutional detection-regression network. Joints and associations between joint pairs in our instrument model are located by the detection subnetwork and are subsequently refined through a regression sub-network. Based on the output from the model, the poses of the instruments are inferred using maximum bipartite graph matching. Our estimation framework is powered by deep learning techniques without any direct kinematic information from a robot. Our framework is tested on single-instrument RMIT data, and also on multi-instrument EndoVis and in vivo data with promising results. In addition, the data set annotations are publicly released along with our code and model.
引用
收藏
页码:1276 / 1287
页数:12
相关论文
共 31 条
[1]
Image Based Surgical Instrument Pose Estimation with Multi-class Labelling and Optical Flow [J].
Allan, Max ;
Chang, Ping-Lin ;
Ourselin, Sebastien ;
Hawkes, David J. ;
Sridhar, Ashwin ;
Kelly, John ;
Stoyanov, Danail .
MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION - MICCAI 2015, PT I, 2015, 9349 :331-338
[2]
Vision-based and marker-less surgical tool detection and tracking: a review of the literature [J].
Bouget, David ;
Allan, Max ;
Stoyanov, Danail ;
Jannin, Pierre .
MEDICAL IMAGE ANALYSIS, 2017, 35 :633-654
[3]
Detecting Surgical Tools by Modelling Local Appearance and Global Shape [J].
Bouget, David ;
Benenson, Rodrigo ;
Omran, Mohamed ;
Riffaud, Laurent ;
Schiele, Bernt ;
Jannin, Pierre .
IEEE TRANSACTIONS ON MEDICAL IMAGING, 2015, 34 (12) :2603-2617
[4]
Human Pose Estimation via Convolutional Part Heatmap Regression [J].
Bulat, Adrian ;
Tzimiropoulos, Georgios .
COMPUTER VISION - ECCV 2016, PT VII, 2016, 9911 :717-732
[5]
Cao Zhe, 2016, REALTIME MULTIPERSON
[6]
Combined 2D and 3D tracking of surgical instruments for minimally invasive and robotic-assisted surgery [J].
Du, Xiaofei ;
Allan, Maximilian ;
Dore, Alessio ;
Ourselin, Sebastien ;
Hawkes, David ;
Kelly, John D. ;
Stoyanov, Danail .
INTERNATIONAL JOURNAL OF COMPUTER ASSISTED RADIOLOGY AND SURGERY, 2016, 11 (06) :1109-1119
[7]
Progressive search space reduction for human pose estimation [J].
Ferrari, Vittorio ;
Marin-Jimenez, Manuel ;
Zisserman, Andrew .
2008 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOLS 1-12, 2008,
[8]
Garcia-Peraza-Herrera L. C., 2017, TOOLNET HOLISTICALLY
[9]
Real-Time Segmentation of Non-rigid Surgical Tools Based on Deep Learning and Tracking [J].
Garcia-Peraza-Herrera, Luis C. ;
Li, Wenqi ;
Gruijthuijsen, Caspar ;
Devreker, Alain ;
Attilakos, George ;
Deprest, Jan ;
Vander Poorten, Emmanuel ;
Stoyanov, Danail ;
Vercauteren, Tom ;
Ourselin, Sebastien .
COMPUTER-ASSISTED AND ROBOTIC ENDOSCOPY, 2017, 10170 :84-95
[10]
Kurmann Thomas, 2017, Medical Image Computing and Computer-Assisted Intervention, MICCAI 2017. 20th International Conference. Proceedings: LNCS 10434, P505, DOI 10.1007/978-3-319-66185-8_57