On the Performance of One-Stage and Two-Stage Object Detectors in Autonomous Vehicles Using Camera Data

被引:115
作者
Carranza-Garcia, Manuel [1 ]
Torres-Mateo, Jesus [1 ]
Lara-Benitez, Pedro [1 ]
Garcia-Gutierrez, Jorge [1 ]
机构
[1] Univ Seville, Div Comp Sci, ES-41012 Seville, Spain
关键词
autonomous vehicles; convolutional neural networks; deep learning; object detection; transfer learning;
D O I
10.3390/rs13010089
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Object detection using remote sensing data is a key task of the perception systems of self-driving vehicles. While many generic deep learning architectures have been proposed for this problem, there is little guidance on their suitability when using them in a particular scenario such as autonomous driving. In this work, we aim to assess the performance of existing 2D detection systems on a multi-class problem (vehicles, pedestrians, and cyclists) with images obtained from the on-board camera sensors of a car. We evaluate several one-stage (RetinaNet, FCOS, and YOLOv3) and two-stage (Faster R-CNN) deep learning meta-architectures under different image resolutions and feature extractors (ResNet, ResNeXt, Res2Net, DarkNet, and MobileNet). These models are trained using transfer learning and compared in terms of both precision and efficiency, with special attention to the real-time requirements of this context. For the experimental study, we use the Waymo Open Dataset, which is the largest existing benchmark. Despite the rising popularity of one-stage detectors, our findings show that two-stage detectors still provide the most robust performance. Faster R-CNN models outperform one-stage detectors in accuracy, being also more reliable in the detection of minority classes. Faster R-CNN Res2Net-101 achieves the best speed/accuracy tradeoff but needs lower resolution images to reach real-time speed. Furthermore, the anchor-free FCOS detector is a slightly faster alternative to RetinaNet, with similar precision and lower memory usage.
引用
收藏
页码:1 / 23
页数:23
相关论文
共 46 条
  • [31] Fully Convolutional Networks for Semantic Segmentation
    Shelhamer, Evan
    Long, Jonathan
    Darrell, Trevor
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (04) : 640 - 651
  • [32] Simonyan A., 2015, P 3 INT C LEARN REPR, P1
  • [33] Scalability in Perception for Autonomous Driving: Waymo Open Dataset
    Sun, Pei
    Kretzschmar, Henrik
    Dotiwalla, Xerxes
    Chouard, Aurelien
    Patnaik, Vijaysai
    Tsui, Paul
    Guo, James
    Zhou, Yin
    Chai, Yuning
    Caine, Benjamin
    Vasudevan, Vijay
    Han, Wei
    Ngiam, Jiquan
    Zhao, Hang
    Timofeev, Aleksei
    Ettinger, Scott
    Krivokon, Maxim
    Gao, Amy
    Joshi, Aditya
    Zhang, Yu
    Shlens, Jonathon
    Chen, Zhifeng
    Anguelov, Dragomir
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 2443 - 2451
  • [34] FCOS: Fully Convolutional One-Stage Object Detection
    Tian, Zhi
    Shen, Chunhua
    Chen, Hao
    He, Tong
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 9626 - 9635
  • [35] Autonomous vehicle perception: The technology of today and tomorrow
    Van Brummelen, Jessica
    O'Brien, Marie
    Gruyer, Dominique
    Najjaran, Homayoun
    [J]. TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2018, 89 : 384 - 406
  • [36] Deep visual domain adaptation: A survey
    Wang, Mei
    Deng, Weihong
    [J]. NEUROCOMPUTING, 2018, 312 : 135 - 153
  • [37] Pseudo-LiDAR from Visual Depth Estimation: Bridging the Gap in 3D Object Detection for Autonomous Driving
    Wang, Yan
    Chao, Wei-Lun
    Garg, Divyansh
    Hariharan, Bharath
    Campbell, Mark
    Weinberger, Kilian Q.
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 8437 - 8445
  • [38] Anchor Generation Optimization and Region of Interest Assignment for Vehicle Detection
    Wang, Ye
    Liu, Zhenyi
    Deng, Weiwen
    [J]. SENSORS, 2019, 19 (05)
  • [39] Wesolowski L, 2018, IEEE I CONF COMP VIS
  • [40] Wu Yuxin, Detectron2