Automatic detection of invasive ductal carcinoma in whole slide images with Convolutional Neural Networks

被引:297
作者
Cruz-Roa, Angel [2 ]
Basavanhally, Ajay [3 ]
Gonzalez, Fabio [2 ]
Gilmore, Hannah [4 ]
Feldman, Michael [5 ]
Ganesan, Shridar [6 ]
Shih, Natalie [5 ]
Tomaszewski, John [7 ]
Madabhushi, Anant [1 ]
机构
[1] Case Western Reserve Univ, Cleveland, OH 44106 USA
[2] Univ Nacl Colombia, Bogota, Colombia
[3] Rutgers State Univ, Piscataway, NJ 08855 USA
[4] Univ Hosp, Cleveland, OH USA
[5] Hosp Univ Penn, Philadelphia, PA 19104 USA
[6] Canc Inst New Jerseey, New Brunswick, NJ USA
[7] SUNY Buffalo, Buffalo, NY USA
来源
MEDICAL IMAGING 2014: DIGITAL PATHOLOGY | 2014年 / 9041卷
基金
美国国家卫生研究院;
关键词
Breast cancer; convolutional neural networks; deep learning; digital pathology; whole-slide imaging; invasive ductal carcinoma; handcrafted features; BREAST-CANCER; PATHOLOGY; COLOR; GRADE;
D O I
10.1117/12.2043872
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
This paper presents a deep learning approach for automatic detection and visual analysis of invasive ductal carcinoma (IDC) tissue regions in whole slide images (WSI) of breast cancer (BCa). Deep learning approaches are learn-from-data methods involving computational modeling of the learning process. This approach is similar to how human brain works using different interpretation levels or layers of most representative and useful features resulting into a hierarchical learned representation. These methods have been shown to outpace traditional approaches of most challenging problems in several areas such as speech recognition and object detection. Invasive breast cancer detection is a time consuming and challenging task primarily because it involves a pathologist scanning large swathes of benign regions to ultimately identify the areas of malignancy. Precise delineation of IDC in WSI is crucial to the subsequent estimation of grading tumor aggressiveness and predicting patient outcome. DL approaches are particularly adept at handling these types of problems, especially if a large number of samples are available for training, which would also ensure the generalizability of the learned features and classifier. The DL framework in this paper extends a number of convolutional neural networks (CNN) for visual semantic analysis of tumor regions for diagnosis support. The CNN is trained over a large amount of image patches (tissue regions) from WSI to learn a hierarchical part-based representation. The method was evaluated over a WSI dataset from 162 patients diagnosed with IDC. 113 slides were selected for training and 49 slides were held out for independent testing. Ground truth for quantitative evaluation was provided via expert delineation of the region of cancer by an expert pathologist on the digitized slides. The experimental evaluation was designed to measure classifier accuracy in detecting IDC tissue regions in WSI. Our method yielded the best quantitative results for automatic detection of IDC regions in WSI in terms of F-measure and balanced accuracy (71.80%, 84.23%), in comparison with an approach using handcrafted image features (color, texture and edges, nuclear textural and architecture), and a machine learning classifier for invasive tumor classification using a Random Forest. The best performing handcrafted features were fuzzy color histogram (67.53%, 78.74%) and RGB histogram (66.64%, 77.24%). Our results also suggest that at least some of the tissue classification mistakes (false positives and false negatives) were less due to any fundamental problems associated with the approach, than the inherent limitations in obtaining a very highly granular annotation of the diseased area of interest by an expert pathologist.
引用
收藏
页数:15
相关论文
共 41 条
[1]   Face description with local binary patterns:: Application to face recognition [J].
Ahonen, Timo ;
Hadid, Abdenour ;
Pietikainen, Matti .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2006, 28 (12) :2037-2041
[2]  
Cruz-Roa AA, 2013, LECT NOTES COMPUT SC, V8150, P403, DOI 10.1007/978-3-642-40763-5_50
[3]  
[Anonymous], 2008, P 16 INT C MULTIMEDI, DOI DOI 10.1145/1459359.1459577
[4]  
[Anonymous], 2010, Advances in Neural Information Processing Systems
[5]   Deep Machine Learning-A New Frontier in Artificial Intelligence Research [J].
Arel, Itamar ;
Rose, Derek C. ;
Karnowski, Thomas P. .
IEEE COMPUTATIONAL INTELLIGENCE MAGAZINE, 2010, 5 (04) :13-18
[6]   Multi-Field-of-View Framework for Distinguishing Tumor Grade in ER plus Breast Cancer From Entire Histopathology Slides [J].
Basavanhally, Ajay ;
Ganesan, Shridar ;
Feldman, Michael ;
Shih, Natalie ;
Mies, Carolyn ;
Tomaszewski, John ;
Madabhushi, Anant .
IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, 2013, 60 (08) :2089-2099
[7]   Representation Learning: A Review and New Perspectives [J].
Bengio, Yoshua ;
Courville, Aaron ;
Vincent, Pascal .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (08) :1798-1828
[8]   Learning Deep Architectures for AI [J].
Bengio, Yoshua .
FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2009, 2 (01) :1-127
[9]  
Boulanger-Lewandowski N., 2012, Modeling temporal dependencies in high-dimensional sequences: Application to polyphonic music generation and transcription
[10]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32