R2-CNN: Fast Tiny Object Detection in Large-Scale Remote Sensing Images

被引:193
作者
Pang, Jiangmiao [1 ,2 ]
Li, Cong [3 ]
Shi, Jianping [3 ]
Xu, Zhihai [1 ,2 ]
Feng, Huajun [1 ,2 ]
机构
[1] Zhejiang Univ, State Key Lab Modern Opt Instrumentat, Hangzhou 310027, Zhejiang, Peoples R China
[2] Zhejiang Univ, Coll Opt Sci & Engn, Hangzhou 310027, Zhejiang, Peoples R China
[3] SenseTime Res, Beijing 100084, Peoples R China
来源
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING | 2019年 / 57卷 / 08期
关键词
Object detection; remote sensing images; remote sensing region-based convolutional neural network (R-2-CNN); ORIENTED GRADIENTS; HISTOGRAMS; NETWORKS;
D O I
10.1109/TGRS.2019.2899955
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
Recently, the convolutional neural network has brought impressive improvements for object detection. However, detecting tiny objects in large-scale remote sensing images still remains challenging. First, the extreme large input size makes the existing object detection solutions too slow for practical use. Second, the massive and complex backgrounds cause serious false alarms. Moreover, the ultratiny objects increase the difficulty of accurate detection. To tackle these problems, we propose a unified and self-reinforced network called remote sensing region-based convolutional neural network (R-2-CNN), composing of backbone Tiny-Net, intermediate global attention block, and final classifier and detector. Tiny-Net is a lightweight residual structure, which enables fast and powerful features extraction from inputs. Global attention block is built upon Tiny-Net to inhibit false positives. Classifier is then used to predict the existence of target in each patch, and detector is followed to locate them accurately if available. The classifier and detector are mutually reinforced with end-to-end training, which further speed up the process and avoid false alarms. Effectiveness of R-2-CNN is validated on hundreds of GF-1 images and GF-2 images that are 18 000 x 18 192 pixels, 2.0-m resolution, and 27 620 x 29 200 pixels, 0.8-m resolution, respectively. Specifically, we can process a GF-1 image in 29.4 s on Titian X just with single thread. According to our knowledge, no previous solution can detect the tiny object on such huge remote sensing images gracefully. We believe that it is a significant step toward practical real-time remote sensing systems.
引用
收藏
页码:5512 / 5524
页数:13
相关论文
共 41 条
  • [1] [Anonymous], 2005, PROC CVPR IEEE
  • [2] VHR Object Detection Based on Structural Feature Extraction and Query Expansion
    Bai, Xiao
    Zhang, Huigang
    Zhou, Jun
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2014, 52 (10): : 6508 - 6520
  • [3] Learning Rotation-Invariant Convolutional Neural Networks for Object Detection in VHR Optical Remote Sensing Images
    Cheng, Gong
    Zhou, Peicheng
    Han, Junwei
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2016, 54 (12): : 7405 - 7415
  • [4] A survey on object detection in optical remote sensing images
    Cheng, Gong
    Han, Junwei
    [J]. ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2016, 117 : 11 - 28
  • [5] Dai J, 2016, PROCEEDINGS 2016 IEEE INTERNATIONAL CONFERENCE ON INDUSTRIAL TECHNOLOGY (ICIT), P1796, DOI 10.1109/ICIT.2016.7475036
  • [6] Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
  • [7] The Pascal Visual Object Classes (VOC) Challenge
    Everingham, Mark
    Van Gool, Luc
    Williams, Christopher K. I.
    Winn, John
    Zisserman, Andrew
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2010, 88 (02) : 303 - 338
  • [8] Girshick R., 2015, P IEEE INT C COMPUTE, DOI [DOI 10.1109/ICCV.2015.169, 10.1109/ICCV.2015.169]
  • [9] Girshick R., 2014, IEEE COMP SOC C COMP, DOI [10.1109/CVPR.2014.81, DOI 10.1109/CVPR.2014.81]
  • [10] Object Detection in Optical Remote Sensing Images Based on Weakly Supervised Learning and High-Level Feature Learning
    Han, Junwei
    Zhang, Dingwen
    Cheng, Gong
    Guo, Lei
    Ren, Jinchang
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2015, 53 (06): : 3325 - 3337