Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification

被引:7936
作者
He, Kaiming
Zhang, Xiangyu
Ren, Shaoqing
Sun, Jian
机构
来源
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV) | 2015年
关键词
D O I
10.1109/ICCV.2015.123
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Rectified activation units (rectifiers) are essential for state-of-the-art neural networks. In this work, we study rectifier neural networks for image classification from two aspects. First, we propose a Parametric Rectified Linear Unit (PReLU) that generalizes the traditional rectified unit. PReLU improves model fitting with nearly zero extra computational cost and little overfitting risk. Second, we derive a robust initialization method that particularly considers the rectifier nonlinearities. This method enables us to train extremely deep rectified models directly from scratch and to investigate deeper or wider network architectures. Based on the learnable activation and advanced initialization, we achieve 4.94% top-5 test error on the ImageNet 2012 classification dataset. This is a 26% relative improvement over the ILSVRC 2014 winner (GoogLeNet, 6.66% [33]). To our knowledge, our result is the first(1) to surpass the reported human-level performance (5.1%, [26]) on this dataset.
引用
收藏
页码:1026 / 1034
页数:9
相关论文
共 38 条
[1]  
Agostinelli F., 2014, arXiv preprint arXiv:1412.6830
[2]  
[Anonymous], 2014, Deeply Supervised Nets
[3]  
[Anonymous], 2015, Very Deep Convolu- tional Networks for Large-Scale Image Recognition
[4]  
[Anonymous], 2014, NIPS
[5]  
[Anonymous], ARXIV14121710
[6]  
[Anonymous], 2013, ARXIV13126120
[7]  
[Anonymous], 2014, EUROPEAN C COMPUTER
[8]  
[Anonymous], 2009, PROC IEEE C COMPUT V
[9]  
[Anonymous], 2013, P 30 INT C MACH LEAR
[10]  
[Anonymous], 2013, P INT C MACHINE LEAR