A new hyperparameters optimization method for convolutional neural networks

被引:64
作者
Cui, Hua [1 ]
Bai, Jie [2 ]
机构
[1] Yulin Normal Univ, 1303 Jiaoyudong Rd, Yulin 537000, Peoples R China
[2] Tongji Univ, 1239 Siping Rd, Shanghai 200092, Peoples R China
关键词
Convolutional neural networks; Hyperparameters optimization; Multilevel evolutionary optimization; Bayesian optimization;
D O I
10.1016/j.patrec.2019.02.009
中图分类号
TP18 [人工智能理论];
学科分类号
140502 [人工智能];
摘要
The use of convolutional neural networks involves hyperparameters optimization. Gaussian process based Bayesian optimization (GPEI) has proven to be an effective algorithm to optimize several hyperparameters. Then deep networks for global optimization algorithm (DNGO) that used neural network as an alternative to Gaussian process was proposed to optimize more hyperparameters. This paper presents a new algorithm that combines multiscale and multilevel evolutionary optimization (MSMLEO) with GPEI to optimize dozens of hyperparameters. These hyperparameters are divided into two groups. The first group related with the sizes of layers and kernels are discrete integers. The second group related with learning rates and so on is continuous floating-point numbers. All combinations of the first group are corresponding to the combinations of grid points on multi-scale grids and MSMLEO launches GPEI to optimize the second group of hyperparameters while the first group keeps fixed. The output of convolutional networks configured with above two groups of optimized hyperparameters is used as the fitness of MSMLEO. MSMLEO alternates with GPEI to search the optimal hyperparameters from coarsest scale to finest scale. Experimental results show that our algorithm has better performance and adaptability on optimizing dozens of hyperparameters of neural networks with a variety of numerical types. (C) 2019 Published by Elsevier B.V.
引用
收藏
页码:828 / 834
页数:7
相关论文
共 10 条
[1]
[Anonymous], 2014, INT C LEARN REPR
[2]
[Anonymous], 2011, INT J APPL ENG RES
[3]
[Anonymous], STATISTICS
[4]
[Anonymous], 1998, POPUL ENVIRON
[5]
Bergstra J, 2012, J MACH LEARN RES, V13, P281
[6]
Hutter Frank, 2011, Learning and Intelligent Optimization. 5th International Conference, LION 5. Selected Papers, P507, DOI 10.1007/978-3-642-25566-3_40
[7]
ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[8]
Mockus J., 1978, Towards global optimisation. II, P117
[9]
Snoek J., 2012, Adv Neural Inf Process Syst, V25
[10]
Snoek J, 2014, PR MACH LEARN RES, V32, P1674