Transferring Deep Convolutional Neural Networks for the Scene Classification of High-Resolution Remote Sensing Imagery

被引:982
作者
Hu, Fan [1 ,2 ]
Xia, Gui-Song [1 ]
Hu, Jingwen [1 ,2 ]
Zhang, Liangpei [1 ]
机构
[1] State Key Lab Informat Engn Surveying Mapping & R, Wuhan 430079, Peoples R China
[2] Wuhan Univ, Elect Informat Sch, Wuhan 430072, Peoples R China
基金
中国国家自然科学基金;
关键词
CNN; scene classification; feature representation; feature coding; convolutional layer; fully-connected layer; COOCCURRENCE; FEATURES; SCALE;
D O I
10.3390/rs71114680
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Learning efficient image representations is at the core of the scene classification task of remote sensing imagery. The existing methods for solving the scene classification task, based on either feature coding approaches with low-level hand-engineered features or unsupervised feature learning, can only generate mid-level image features with limited representative ability, which essentially prevents them from achieving better performance. Recently, the deep convolutional neural networks (CNNs), which are hierarchical architectures trained on large-scale datasets, have shown astounding performance in object recognition and detection. However, it is still not clear how to use these deep convolutional neural networks for high-resolution remote sensing (HRRS) scene classification. In this paper, we investigate how to transfer features from these successfully pre-trained CNNs for HRRS scene classification. We propose two scenarios for generating image features via extracting CNN features from different layers. In the first scenario, the activation vectors extracted from fully-connected layers are regarded as the final image features; in the second scenario, we extract dense features from the last convolutional layer at multiple scales and then encode the dense features into global image features through commonly used feature coding approaches. Extensive experiments on two public scene classification datasets demonstrate that the image features obtained by the two proposed scenarios, even with a simple linear classifier, can result in remarkable performance and improve the state-of-the-art by a significant margin. The results reveal that the features from pre-trained CNNs generalize well to HRRS datasets and are more expressive than the low- and mid-level features. Moreover, we tentatively combine features extracted from different CNN models for better performance.
引用
收藏
页码:14680 / 14707
页数:28
相关论文
共 59 条
[21]  
Jaiswal R.K., 1999, J INDIAN SOC REMOTE, V27, P123, DOI DOI 10.1007/BF02990808
[22]  
Jégou H, 2010, PROC CVPR IEEE, P3304, DOI 10.1109/CVPR.2010.5540039
[23]   Caffe: Convolutional Architecture for Fast Feature Embedding [J].
Jia, Yangqing ;
Shelhamer, Evan ;
Donahue, Jeff ;
Karayev, Sergey ;
Long, Jonathan ;
Girshick, Ross ;
Guadarrama, Sergio ;
Darrell, Trevor .
PROCEEDINGS OF THE 2014 ACM CONFERENCE ON MULTIMEDIA (MM'14), 2014, :675-678
[24]   Dirichlet-based Histogram Feature Transform for Image Classification [J].
Kobayashi, Takumi .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :3278-3285
[25]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[26]  
Lazebnik S., 2006, P IEEE INT C COMP VI, V2, P2169, DOI [DOI 10.1109/CVPR.2006.68, 10.1109/CVPR.2006.68]
[27]   Gradient-based learning applied to document recognition [J].
Lecun, Y ;
Bottou, L ;
Bengio, Y ;
Haffner, P .
PROCEEDINGS OF THE IEEE, 1998, 86 (11) :2278-2324
[28]   TEXTURE ANALYSIS WITH SHAPE CO-OCCURRENCE PATTERNS [J].
Liu, Gang ;
Xia, Gui-Song ;
Yang, Wen ;
Zhang, Liangpei .
2014 22ND INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2014, :1627-1632
[29]   Distinctive image features from scale-invariant keypoints [J].
Lowe, DG .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2004, 60 (02) :91-110
[30]  
Mahendran A, 2015, PROC CVPR IEEE, P5188, DOI 10.1109/CVPR.2015.7299155