Comparing humans and deep learning performance for grading AMD: A study in using universal deep features and transfer learning for automated AMD analysis

被引:154
作者
Burlina, Philippe [1 ,2 ,3 ]
Pacheco, Katia D. [4 ]
Joshi, Neil [1 ]
Freund, David E. [1 ]
Bressler, Neil M. [2 ]
机构
[1] Johns Hopkins Univ, Appl Phys Lab, Baltimore, MD 21218 USA
[2] Johns Hopkins Univ, Sch Med, Wilmer Eye Inst, Retina Div, Baltimore, MD 21218 USA
[3] Johns Hopkins Univ, Dept Comp Sci, Baltimore, MD 21218 USA
[4] Vis Eye Hosp, Brazilian Ctr, Retina Div, Brasilia, DF, Brazil
基金
美国国家卫生研究院;
关键词
Deep learning; Deep Convolutional Neural Networks; (DCNNs); Universal features; Retinal image analysis; Age -related macular degeneration; (AMD); Transfer learning;
D O I
10.1016/j.compbiomed.2017.01.018
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Background: When left untreated, age-related macular degeneration (AMD) is the leading cause of vision loss in people over fifty in the US. Currently it is estimated that about eight million US individuals have the intermediate stage of AMD that is often asymptomatic with regard to visual deficit. These individuals are at high risk for progressing to the advanced stage where the often treatable choroidal neovascular form of AMD can occur. Careful monitoring to detect the onset and prompt treatment of the neovascular form as well as dietary supplementation can reduce the risk of vision loss from AMD, therefore, preferred practice patterns recommend identifying individuals with the intermediate stage in a timely manner. Methods: Past automated retinal image analysis (ARIA) methods applied on fundus imagery have relied on engineered and hand-designed visual features. We instead detail the novel application of a machine learning approach using deep learning for the problem of ARIA and AMD analysis. We use transfer learning and universal features derived from deep convolutional neural networks (DCNN). We address clinically relevant 4 class, 3-class, and 2-class AMD severity classification problems. Results: Using 5664 color fundus images from the NIH AREDS dataset and DCNN universal features, we obtain values for accuracy for the (4-, 3-, 2-) class classification problem of (79.4%, 81.5%, 93.4%) for machine vs. (75.8%, 85.0%, 95.2%) for physician grading. Discussion: This study demonstrates the efficacy of machine grading based on deep universal features/transfer learning when applied to ARIA and is a promising step in providing a pre-screener to identify individuals with intermediate AMD and also as a tool that can facilitate identifying such individuals for clinical studies aimed at developing improved therapies. It also demonstrates comparable performance between computer and physician grading.
引用
收藏
页码:80 / 86
页数:7
相关论文
共 35 条
[1]  
Abramoff Michael D, 2010, IEEE Rev Biomed Eng, V3, P169, DOI 10.1109/RBME.2010.2084567
[2]  
Age-Related Eye Dis Study Res Grp, 2001, AM J OPHTHALMOL, V132, P668
[3]  
[Anonymous], AM J OPHTHALMOL
[4]  
[Anonymous], 2014, 14091556 ARXIV
[5]  
[Anonymous], 2007, COHENS KAPPA COMPUTE
[6]  
[Anonymous], 2012, ADV NEURAL INFORM PR
[7]  
[Anonymous], 2014, P IEEE C COMP VIS PA
[8]  
[Anonymous], 2016, I S BIOMED IMAGING, DOI DOI 10.1109/ISBI.2016.7493240
[9]  
[Anonymous], IEEE T PATTERN ANAL
[10]  
[Anonymous], 150602640 ARXIV