Fusion object detection of satellite imagery with arbitrary-oriented region convolutional neural network

被引:3
作者
Ya Y. [1 ]
Pan H. [1 ]
Jing Z. [1 ]
Ren X. [1 ]
Qiao L. [1 ]
机构
[1] School of Aeronautics and Astronautics, Shanghai Jiao Tong University, Shanghai
基金
中国国家自然科学基金;
关键词
CNN; Image fusion; Object detection; Satellite imagery;
D O I
10.1007/s42401-019-00033-x
中图分类号
学科分类号
摘要
Object detection on multi-source images from satellite platforms is difficult due to the characteristics of imaging sensors. Multi-model image fusion provides a possibility to improve the performance of object detection. This paper proposes a fusion object detection framework with arbitrary-oriented region convolutional neural network. First, nine kinds of pansharpening methods are utilized to fuse multi-source images. Second, a novel object detection framework based on Faster Region-based Convolutional Neural Network structure is used, which is suitable for large-scale satellite images. Region Proposal Network is adopted to generate axially aligned bounding boxes enclosing object sin different orientations, and then extract features by pooling layers with different sizes. These features are used to classify the proposals, adjust the bounding boxes, and predict the inclined boxes and the objectness/non-objectness score. Smaller anchors for small objects are considered. Finally, inclined non-maximum suppression method is utilized to get the detection results. Experimental results showed that the proposed method performs better than some state-of-the-art object detection techniques, such as YOLO-v2, YOLO-v3, etc. Some numerical tests validate the efficiency and effectiveness of the proposed method. © 2019, Shanghai Jiao Tong University.
引用
收藏
页码:163 / 174
页数:11
相关论文
共 55 条
  • [1] Krizhevsky A., Sutskever I., Hinton G., Imagenet Classification with Deep Convolutional Neural Networks, (2012)
  • [2] Russakovsky O., Deng J., Su H., Et al., ImageNet large scale visual recognition challenge, Int J Comput Vis, 115, 3, pp. 211-252, (2015)
  • [3] Everingham M., Winn J., The PASCAL visual object classes challenge 2007 (VOC2007) development kit, Int J Comput Vis, 111, 1, pp. 98-136, (2006)
  • [4] Hinton G.E., Salakhutdinov R.R., Reducing the dimensionality of data with neural networks, Science, 313, 5786, pp. 504-507, (2006)
  • [5] LeCun Y., Boser B., Denker J.S., Henderson D., Howard R.E., Hubbard W., Jackel L.D., Backpropagation applied to handwritten zip code recognition, Neural Comput, 1, 4, pp. 541-551, (1989)
  • [6] Krizhevsky A., Sutskever I., Hinton G.E., ImageNet classification with deep convolutional neural networks, Proceedings of the 25Th International Conference on Neural Information Processing Systems, pp. 1097-1105, (2012)
  • [7] Vishwakarma S., Agrawal A., A survey on activity recognition and behavior understanding in video surveillance, Visual Comput, 29, 10, pp. 983-1009, (2013)
  • [8] Zhao Z.Q., Zheng P., Xu S.T., Et al., Object Detection with Deep Learning: A Review, (2018)
  • [9] LeCun Y., Bengio Y., Hinton G., Deep learning, Nature, 521, 7553, pp. 436-444, (2015)
  • [10] Ren S., He K., Girshick R., Et al., Faster R-CNN: towards real-time object detection with region proposal networks, IEEE Trans Pattern Anal Mach Intell, 39, 6, pp. 1137-1149, (2015)