A Deep-Local-Global Feature Fusion Framework for High Spatial Resolution Imagery Scene Classification

被引:59
作者
Zhu, Qiqi [1 ]
Zhong, Yanfei [1 ]
Liu, Yanfei [1 ]
Zhang, Liangpei [1 ]
Li, Deren [1 ]
机构
[1] Wuhan Univ, State Key Lab Informat Engn Surveying Mapping & R, Wuhan 430079, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
scene classification; deep feature; global low-level features; local feature; BoVW; high spatial resolution image; fusion; REMOTE; SCALE; MODEL;
D O I
10.3390/rs10040568
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
High spatial resolution (HSR) imagery scene classification has recently attracted increased attention. The bag-of-visual-words (BoVW) model is an effective method for scene classification. However, it can only extract handcrafted features, and it disregards the spatial layout information, whereas deep learning can automatically mine the intrinsic features as well as preserve the spatial location, but it may lose the characteristic information of the HSR images. Although previous methods based on the combination of BoVW and deep learning have achieved comparatively high classification accuracies, they have not explored the combination of handcrafted and deep features, and they just used the BoVW model as a feature coding method to encode the deep features. This means that the intrinsic characteristics of these models were not combined in the previous works. In this paper, to discover more discriminative semantics for HSR imagery, the deep-local-global feature fusion (DLGFF) framework is proposed for HSR imagery scene classification. Differing from the conventional scene classification methods, which utilize only handcrafted features or deep features, DLGFF establishes a framework integrating multi-level semantics from the global texture feature-based method, the BoVW model, and a pre-trained convolutional neural network (CNN). In DLGFF, two different approaches are proposed, i. e., the local and global features fused with the pooling-stretched convolutional features (LGCF) and the local and global features fused with the fully connected features (LGFF), to exploit the multi-level semantics for complex scenes. The experimental results obtained with three HSR image classification datasets confirm the effectiveness of the proposed DLGFF framework. Compared with the published results of the previous scene classification methods, the classification accuracies of the DLGFF framework on the 21-class UC Merced dataset and 12-class Google dataset of SIRI-WHU can reach 99.76%, which is superior to the current state-of-the-art methods. The classification accuracy of the DLGFF framework on the 45-class NWPU-RESISC45 dataset, 96.37 +-0.05%, is an increase of about 6% when compared with the current state-of-the-art methods. This indicates that the fusion of the global low-level feature, the local mid-level feature, and the deep high-level feature can provide a representative description for HSR imagery.
引用
收藏
页数:22
相关论文
共 47 条
[1]  
[Anonymous], 2013, IEEE T PATTERN ANAL, DOI DOI 10.1109/TPAMI.2012.59
[2]  
[Anonymous], ATONMET DEEP CONVOLU
[3]  
[Anonymous], P INT C IM PROC BARC
[4]  
[Anonymous], PROC CVPR IEEE
[5]   Geographic Object-Based Image Analysis - Towards a new paradigm [J].
Blaschke, Thomas ;
Hay, Geoffrey J. ;
Kelly, Maggi ;
Lang, Stefan ;
Hofmann, Peter ;
Addink, Elisabeth ;
Feitosa, Raul Queiroz ;
van der Meer, Freek ;
van der Werff, Harald ;
van Coillie, Frieke ;
Tiede, Dirk .
ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2014, 87 :180-191
[6]  
Boureau Y. L., 2010, P ICML 10 P 27 INT C, P111
[7]   Bridging the Semantic Gap for Satellite Image Annotation and Automatic Mapping Applications [J].
Bratasanu, Dragos ;
Nedelcu, Ion ;
Datcu, Mihai .
IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2011, 4 (01) :193-204
[8]  
Castelluccio M., 2015, Land Use Classification in Remote Sensing Images by Convolutional Neural Networks
[9]   Pyramid of Spatial Relatons for Scene-Level Land Use Classification [J].
Chen, Shizhi ;
Tian, YingLi .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2015, 53 (04) :1947-1957
[10]  
Cheng G., 2017, P IEEE, V105, P1865, DOI DOI 10.1109/JPROC.2017.2675998