Representing and reducing error in natural-resource classification using model combination

被引:18
作者
Huang, Z [1 ]
Lees, BG [1 ]
机构
[1] Australian Natl Univ, Sch Resources Environm & Soc, Canberra, ACT 0200, Australia
关键词
representing error; reducing error; natural-resource classification; model combination;
D O I
10.1080/13658810500032446
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Artificial Intelligence ( AI) models such as Artificial Neural Networks (ANNs), Decision Trees and Dempster-Shafer's Theory of Evidence have long claimed to be more error-tolerant than conventional statistical models, but the way error is propagated through these models is unclear. Two sources of error have been identified in this study: sampling error and attribute error. The results show that these errors propagate differently through the three AI models. The Decision Tree was the most affected by error, the Artificial Neural Network was less affected by error, and the Theory of Evidence model was not affected by the errors at all. The study indicates that AI models have very different modes of handling errors. In this case, the machine-learning models, including ANNs and Decision Trees, are more sensitive to input errors. Dempster-Shafer's Theory of Evidence has demonstrated better potential in dealing with input errors when multisource data sets are involved. The study suggests a strategy of combining AI models to improve classification accuracy. Several combination approaches have been applied, based on a 'majority voting system', a simple average, Dempster Shafer's Theory of Evidence, and fuzzy-set theory. These approaches all increased classification accuracy to some extent. Two of them also demonstrated good performance in handling input errors. Second-stage combination approaches which use statistical evaluation of the initial combinations are able to further improve classification results. One of these second-stage combination approaches increased the overall classification accuracy on forest types to 54% from the original 46.5% of the Decision Tree model, and its visual appearance is also much closer to the ground data. By combining models, it becomes possible to calculate quantitative confidence measurements for the classification results, which can then serve as a better error representation. Final classification products include not only the predicted hard classes for individual cells, but also estimates of the probability and the confidence measurements of the prediction.
引用
收藏
页码:603 / 621
页数:19
相关论文
共 38 条
[1]  
ABRAHART RJ, 1998, P 3 INT C GEOCOMPUTA
[2]   MULTISPECTRAL CLASSIFICATION OF LANDSAT-IMAGES USING NEURAL NETWORKS [J].
BISCHOF, H ;
SCHNEIDER, W ;
PINZ, AJ .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 1992, 30 (03) :482-490
[3]  
DRUMMOND J, 1987, ITC J, V1, P73
[4]  
DRUMMOND J, 1996, SPAT ACC ASS NAT RES, P189
[5]   INTEGRATION OF MULTI-AI PARADIGMS FOR INTELLIGENT OPERATION SUPPORT SYSTEMS - FUZZY RULE EXTRACTION FROM A NEURAL-NETWORK [J].
ENBUTSU, I ;
BABA, K ;
HARA, N ;
WASEDA, K ;
NOGITA, S .
WATER SCIENCE AND TECHNOLOGY, 1993, 28 (11-12) :333-340
[6]  
FITZGERALD RW, 1994, ANU KIOLOA 1994 REMO
[7]  
FORIER F, 1996, SPATIAL ACCURACY ASS, P225
[8]   STOCHASTIC RELAXATION, GIBBS DISTRIBUTIONS, AND THE BAYESIAN RESTORATION OF IMAGES [J].
GEMAN, S ;
GEMAN, D .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1984, 6 (06) :721-741
[9]  
GOODCHILD MF, 1989, ACCURACY OF SPATIAL DATABASES, P107
[10]   DEVELOPMENT AND TEST OF AN ERROR MODEL FOR CATEGORICAL-DATA [J].
GOODCHILD, MF ;
SUN, GQ ;
SHIREN, Y .
INTERNATIONAL JOURNAL OF GEOGRAPHICAL INFORMATION SYSTEMS, 1992, 6 (02) :87-104