A Comparison of Prediction Accuracy, Complexity, and Training Time of Thirty-Three Old and New Classification Algorithms

被引:1
作者
Tjen-Sien Lim
Wei-Yin Loh
Yu-Shan Shih
机构
[1] University of Wisconsin,Department of Statistics
[2] University of Wisconsin,Department of Statistics
[3] National Chung Cheng University,Department of Mathematics
来源
Machine Learning | 2000年 / 40卷
关键词
classification tree; decision tree; neural net; statistical classifier;
D O I
暂无
中图分类号
学科分类号
摘要
Twenty-two decision tree, nine statistical, and two neural network algorithms are compared on thirty-two datasets in terms of classification accuracy, training time, and (in the case of trees) number of leaves. Classification accuracy is measured by mean error rate and mean rank of error rate. Both criteria place a statistical, spline-based, algorithm called POLYCLSSS at the top, although it is not statistically significantly different from twenty other algorithms. Another statistical algorithm, logistic regression, is second with respect to the two accuracy criteria. The most accurate decision tree algorithm is QUEST with linear splits, which ranks fourth and fifth, respectively. Although spline-based statistical algorithms tend to have good accuracy, they also require relatively long training times. POLYCLASS, for example, is third last in terms of median training time. It often requires hours of training compared to seconds for other algorithms. The QUEST and logistic regression algorithms are substantially faster. Among decision tree algorithms with univariate splits, C4.5, IND-CART, and QUEST have the best combinations of error rate and speed. But C4.5 tends to produce trees with twice as many leaves as those from IND-CART and QUEST.
引用
收藏
页码:203 / 228
页数:25
相关论文
共 51 条
[1]  
Breslow L. A.(1997)Simplifying decision trees: A survey Knowledge Engineering Review 12 1-40
[2]  
Aha D. W.(1995)Multivariate decision trees Machine Learning 19 45-77
[3]  
Brodley C. E.(1993)A comparison of decision tree classifiers with backpropagation neural networks for multimodal classification problems Pattern Recognition 26 953-961
[4]  
Utgoff P. E.(1992)Learning classification trees Statistics and Computing 2 63-73
[5]  
Brown D. E.(1994)Neural networks, decision tree induction and discriminant analysis: An empirical comparison Journal of the Operational Research Society 45 440-450
[6]  
Corruble V.(1991)Multivariate adaptive regression splines (with discussion) Annals of Statistics 19 1-141
[7]  
Pittard C. L.(1937)The use of ranks to avoid the assumption of normality implicit in the analysis of variance Journal of the American Statistical Association 32 675-701
[8]  
Buntine W.(1978)Hedonic prices and the demand for clean air Journal of Environmental Economics and Management 5 81-102
[9]  
Curram S. P.(1995)Penalized discriminant analysis Annals of Statistics 23 73-102
[10]  
Mingers J.(1996)Discriminant analysis by Gaussian mixtures Journal of the Royal Statistical Society, Series B 58 155-176