Automated detection of workers and heavy equipment on construction sites: A convolutional neural network approach

被引:246
作者
Fang, Weili [1 ,2 ]
Ding, Lieyun [1 ,2 ]
Zhong, Botao [1 ,2 ]
Love, Peter E. D. [3 ]
Luo, Hanbin [1 ,2 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Civil Engn & Mech, Wuhan 430074, Hubei, Peoples R China
[2] HUST, Hubei Engn Res Ctr Virtual Safe & Automated Const, Wuhan, Hubei, Peoples R China
[3] Curtin Univ, Dept Civil Engn, Perth, WA 6023, Australia
基金
中国国家自然科学基金;
关键词
Deep learning; Image; Improved Faster R-CNN; Object detection; Construction site; ORIENTED GRADIENTS; RECOGNITION; HISTOGRAMS; IMAGES; MODEL;
D O I
10.1016/j.aei.2018.05.003
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Detecting the presence of workers, plant, equipment, and materials (i.e. objects) on sites to improve safety and productivity has formed an integral part of computer vision-based research in construction. Such research has tended to focus on the use of computer vision and pattern recognition approaches that are overly reliant on the manual extraction of features and small datasets (< 10k images/label), which can limit inter and intra-class variability. As a result, this hinders their ability to accurately detect objects on construction sites and generalization to different datasets. To address this limitation, an Improved Faster Regions with Convolutional Neural Network Features (IFaster R-CNN) approach is used to automatically detect the presence of objects in real-time is developed, which comprises: (1) the establishment dataset of workers and heavy equipment to train the CNN; (2) extraction of feature maps from images using deep model; (3) extraction of a region proposal from feature maps; and (4) object recognition. To validate the model's ability to detect objects in real-time, a specific dataset is established to train the Waster R-CNN models to detect workers and plant (e.g. excavator). The results reveal that the Waster R-CNN is able to detect the presence of workers and excavators at a high level of accuracy (91% and 95%). The accuracy of the proposed deep learning method exceeds that of current state-of-the-art descriptor methods in detecting target objects on images.
引用
收藏
页码:139 / 149
页数:11
相关论文
共 49 条
[1]   Automated Visual Recognition of Dump Trucks in Construction Videos [J].
Azar, Ehsan Rezazadeh ;
McCabe, Brenda .
JOURNAL OF COMPUTING IN CIVIL ENGINEERING, 2012, 26 (06) :769-781
[2]   Part based model and spatial-temporal reasoning to recognize hydraulic excavators in construction images and videos [J].
Azar, Ehsan Rezazadeh ;
McCabe, Brenda .
AUTOMATION IN CONSTRUCTION, 2012, 24 :194-202
[3]   SURF: Speeded up robust features [J].
Bay, Herbert ;
Tuytelaars, Tinne ;
Van Gool, Luc .
COMPUTER VISION - ECCV 2006 , PT 1, PROCEEDINGS, 2006, 3951 :404-417
[4]   Progressive 3D reconstruction of infrastructure with videogrammetry [J].
Brilakis, Ioannis ;
Fathi, Habib ;
Rashidi, Abbas .
AUTOMATION IN CONSTRUCTION, 2011, 20 (07) :884-895
[5]   Deep Learning-Based Crack Damage Detection Using Convolutional Neural Networks [J].
Cha, Young-Jin ;
Choi, Wooram ;
Buyukozturk, Oral .
COMPUTER-AIDED CIVIL AND INFRASTRUCTURE ENGINEERING, 2017, 32 (05) :361-378
[6]   Automated Object Identification Using Optical Video Cameras on Construction Sites [J].
Chi, Seokho ;
Caldas, Carlos H. .
COMPUTER-AIDED CIVIL AND INFRASTRUCTURE ENGINEERING, 2011, 26 (05) :368-380
[7]   Integrating mobile Building Information Modelling and Augmented Reality systems: An experimental study [J].
Chu, Michael ;
Matthews, Jane ;
Love, Peter E. D. .
AUTOMATION IN CONSTRUCTION, 2018, 85 :305-316
[8]   Histograms of oriented gradients for human detection [J].
Dalal, N ;
Triggs, B .
2005 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 1, PROCEEDINGS, 2005, :886-893
[9]   A deep hybrid learning model to detect unsafe behavior: Integrating convolution neural networks and long short-term memory [J].
Ding, Lieyun ;
Fang, Weili ;
Luo, Hanbin ;
Love, Peter E. D. ;
Zhong, Botao ;
Ouyang, Xi .
AUTOMATION IN CONSTRUCTION, 2018, 86 :118-124
[10]   Detecting non-hardhat-use by a deep learning method from far -field surveillance videos [J].
Fang, Qi ;
Li, Heng ;
Luo, Xiaochun ;
Ding, Lieyun ;
Luo, Hanbin ;
Rose, Timothy M. ;
An, Wangpeng .
AUTOMATION IN CONSTRUCTION, 2018, 85 :1-9