Fusing information from multiple 2D depth cameras for 3D human pose estimation in the operating room

被引:21
作者
Hansen, Lasse [1 ]
Siebert, Marlin [1 ]
Diesel, Jasper [2 ]
Heinrich, Mattias P. [1 ]
机构
[1] Univ Lubeck, Inst Med Informat, Ratzeburger Allee 160, D-23562 Lubeck, Germany
[2] Dragerwerk AG & Co KGaA, Moislinger Allee 53-55, D-23558 Lubeck, Germany
关键词
Human pose estimation; Deep learning; 2D-3D information fusion; Convolutional autoencoder; Operating room; RECOGNITION;
D O I
10.1007/s11548-019-02044-7
中图分类号
R318 [生物医学工程];
学科分类号
100103 [病原生物学];
摘要
Purpose For many years, deep convolutional neural networks have achieved state-of-the-art results on a wide variety of computer vision tasks. 3D human pose estimation makes no exception and results on public benchmarks are impressive. However, specialized domains, such as operating rooms, pose additional challenges. Clinical settings include severe occlusions, clutter and difficult lighting conditions. Privacy concerns of patients and staff make it necessary to use unidentifiable data. In this work, we aim to bring robust human pose estimation to the clinical domain. Methods We propose a 2D-3D information fusion framework that makes use of a network of multiple depth cameras and strong pose priors. In a first step, probabilities of 2D joints are predicted from single depth images. These information are fused in a shared voxel space yielding a rough estimate of the 3D pose. Final joint positions are obtained by regressing into the latent pose space of a pre-trained convolutional autoencoder. Results We evaluate our approach against several baselines on the challenging MVOR dataset. Best results are obtained when fusing 2D information from multiple views and constraining the predictions with learned pose priors. Conclusions We present a robust 3D human pose estimation framework based on a multi-depth camera network in the operating room. Depth images as only input modalities make our approach especially interesting for clinical applications due to the given anonymity for patients and staff.
引用
收藏
页码:1871 / 1879
页数:9
相关论文
共 40 条
[1]
Achilles Felix, 2016, Medical Image Computing and Computer-Assisted Intervention - MICCAI 2016. 19th International Conference. Proceedings: LNCS 9900, P491, DOI 10.1007/978-3-319-46720-7_57
[2]
PoseTrack: A Benchmark for Human Pose Estimation and Tracking [J].
Andriluka, Mykhaylo ;
Iqbal, Umar ;
Insafutdinov, Eldar ;
Pishchulin, Leonid ;
Milan, Anton ;
Gall, Juergen ;
Schiele, Bernt .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :5167-5176
[3]
2D Human Pose Estimation: New Benchmark and State of the Art Analysis [J].
Andriluka, Mykhaylo ;
Pishchulin, Leonid ;
Gehler, Peter ;
Schiele, Bernt .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :3686-3693
[4]
Andriluka M, 2009, PROC CVPR IEEE, P1014, DOI 10.1109/CVPRW.2009.5206754
[5]
Parsing human skeletons in an operating room [J].
Belagiannis, Vasileios ;
Wang, Xinchao ;
Ben Shitrit, Horesh Beny ;
Hashimoto, Kiyoshi ;
Stauder, Ralf ;
Aoki, Yoshimitsu ;
Kranzfelder, Michael ;
Schneider, Armin ;
Fua, Pascal ;
Ilic, Slobodan ;
Feussner, Hubertus ;
Navab, Nassir .
MACHINE VISION AND APPLICATIONS, 2016, 27 (07) :1035-1046
[6]
Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields [J].
Cao, Zhe ;
Simon, Tomas ;
Wei, Shih-En ;
Sheikh, Yaser .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1302-1310
[7]
Patient-Specific Pose Estimation in Clinical Environments [J].
Chen, Kenny ;
Gabriel, Paolo ;
Alasfour, Abdulwahab ;
Gong, Chenghao ;
Doyle, Werner K. ;
Devinsky, Orrin ;
Friedman, Daniel ;
Dugan, Patricia ;
Melloni, Lucia ;
Thesen, Thomas ;
Gonda, David ;
Sattar, Shifteh ;
Wang, Sonya ;
Gilja, Vikash .
IEEE JOURNAL OF TRANSLATIONAL ENGINEERING IN HEALTH AND MEDICINE, 2018, 6
[8]
Cascaded Pyramid Network for Multi-Person Pose Estimation [J].
Chen, Yilun ;
Wang, Zhicheng ;
Peng, Yuxiang ;
Zhang, Zhiqiang ;
Yu, Gang ;
Sun, Jian .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7103-7112
[9]
Dietz A., 2016, GCAI, P138
[10]
Felzenszwalb P, 2008, PROC CVPR IEEE, P1984