ImageNet Classification with Deep Convolutional Neural Networks

被引:76294
作者
Krizhevsky, Alex [1 ]
Sutskever, Ilya [2 ]
Hinton, Geoffrey E. [1 ]
机构
[1] Google Inc, Mountain View, CA 94043 USA
[2] OpenAI, San Francisco, CA USA
关键词
33;
D O I
10.1145/3065386
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
We trained a large, deep convolutional neural network to classify the 1.2 million high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 different classes. On the test data, we achieved top-1 and top-5 error rates of 37.5% and 17.0%, respectively, which is considerably better than the previous state-of-the-art. The neural network, which has 60 million parameters and 650,000 neurons, consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully connected layers with a final 1000-way softmax. To make training faster, we used non-saturating neurons and a very efficient GPU implementation of the convolution operation. To reduce overfitting in the fully connected layers we employed a recently developed regularization method called "dropout" that proved to be very effective. We also entered a variant of this model in the ILSVRC-2012 competition and achieved a winning top-5 test error rate of 15.3%, compared to 26.2% achieved by the second-best entry.
引用
收藏
页码:84 / 90
页数:7
相关论文
共 33 条
  • [1] [Anonymous], ECCV EUR C COMP VIS
  • [2] [Anonymous], 2011, ESANN
  • [3] [Anonymous], 2009, CVPR09
  • [4] [Anonymous], 1989, Neural Information Processing Systems
  • [5] [Anonymous], 2011, arXiv preprint arXiv:1102.0183
  • [6] [Anonymous], 1985, PARALLEL DISTRIBUTED
  • [7] [Anonymous], 2007, Acm Sigkdd Explorations Newsletter
  • [8] [Anonymous], 1985, P COGNITIVA
  • [9] [Anonymous], 1974, REGRESSION NEW TOOLS
  • [10] [Anonymous], 2009, ICML