Generalisation for neural networks through data sampling and training procedures, with applications to streamflow predictions

被引:66
作者
Anctil, F
Lauzon, N
机构
[1] Univ Laval, Dept Civil Engn, Quebec City, PQ G1K 7P4, Canada
[2] Golder Assoc Ltd, Calgary, AB T2P 3T1, Canada
关键词
neural networks; generalisation; stacking; bagging; boosting; stop-training; Bayesian regularisation; streamflow modelling;
D O I
10.5194/hess-8-940-2004
中图分类号
P [天文学、地球科学];
学科分类号
07 ;
摘要
Since the 1990s, neural networks have been applied to many studies in hydrology and water resources. Extensive reviews on neural network modelling have identified the major issues affecting modelling performance: one of the most important is generalisation, which refers to building models that can infer the behaviour of the system under study for conditions represented not only in the data employed for training and testing but also for those conditions not present in the data sets but inherent to the system. This work compares five generalisation approaches: stop training, Bayesian regularisation, stacking, bagging and boosting. All have been tested with neural networks in various scientific domains stop training and stacking having been applied regularly in hydrology and water resources for some years, while Bayesian regularisation, bagging and boosting have been less common. The comparison is applied to streamflow modelling with multi-layer perceptron neural networks and the Levenberg-Marquardt algorithm as training procedure. Six catchments, with diverse hydrological behaviours, are employed as test cases to draw general conclusions and guidelines on the use of the generalisation techniques for practitioners in hydrology and water resources. All generalisation approaches provide improved performance compared with standard neural networks without generalisation. Stacking. bagging and boosting, which affect the construction of training sets, provide the best improvement from standard models, compared with stop-training and Bayesian regularisation, which regulate the training algorithm. Stacking performs better than the others although the benefit in performance is slight compared with bagging and boosting; furthermore, it is not consistent from one catchment to another. For a good combination of improvement and stability in modelling performance, the joint use of stop training or Bayesian regularisation with either bagging or boosting is recommended.
引用
收藏
页码:940 / 958
页数:19
相关论文
共 75 条
[21]   Distributed learning with bagging-like performance [J].
Chawla, NV ;
Moore, TE ;
Hall, LO ;
Bowyer, KW ;
Kegelmeyer, WP ;
Springer, C .
PATTERN RECOGNITION LETTERS, 2003, 24 (1-3) :455-471
[22]   Hydrological forecasting with artificial neural networks:: The state of the art [J].
Coulibaly, P ;
Anctil, F ;
Bobée, B .
CANADIAN JOURNAL OF CIVIL ENGINEERING, 1999, 26 (03) :293-304
[23]   Daily reservoir inflow forecasting using artificial neural networks with stopped training approach [J].
Coulibaly, P ;
Anctil, F ;
Bobée, B .
JOURNAL OF HYDROLOGY, 2000, 230 (3-4) :244-257
[24]   Artificial neural network modeling of water table depth fluctuations [J].
Coulibaly, P ;
Anctil, F ;
Aravena, R ;
Bobée, B .
WATER RESOURCES RESEARCH, 2001, 37 (04) :885-896
[25]   Stability problems with artificial neural networks and the ensemble solution [J].
Cunningham, P ;
Carney, J ;
Jacob, S .
ARTIFICIAL INTELLIGENCE IN MEDICINE, 2000, 20 (03) :217-225
[26]   Hydrological modelling using artificial neural networks [J].
Dawson, CW ;
Wilby, RL .
PROGRESS IN PHYSICAL GEOGRAPHY-EARTH AND ENVIRONMENT, 2001, 25 (01) :80-108
[27]   Effect of pruning and early stopping on performance of a boosting ensemble [J].
Drucker, H .
COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2002, 38 (04) :393-406
[28]  
Foresee FD, 1997, 1997 IEEE INTERNATIONAL CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, P1930, DOI 10.1109/ICNN.1997.614194
[29]   A decision-theoretic generalization of on-line learning and an application to boosting [J].
Freund, Y ;
Schapire, RE .
JOURNAL OF COMPUTER AND SYSTEM SCIENCES, 1997, 55 (01) :119-139
[30]   NEURAL NETWORKS AND THE BIAS VARIANCE DILEMMA [J].
GEMAN, S ;
BIENENSTOCK, E ;
DOURSAT, R .
NEURAL COMPUTATION, 1992, 4 (01) :1-58