Bayesian classification and feature reduction using uniform Dirichlet priors

被引:11
作者
Lynch, RS [1 ]
Willett, PK
机构
[1] USN, Undersea Warfare Ctr, Newport, RI 02841 USA
[2] Univ Connecticut, Storrs, CT 06269 USA
来源
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS | 2003年 / 33卷 / 03期
关键词
class-specific features; discrete features; feature selection; neural networks; noninformative prior; UCI repository;
D O I
10.1109/TSMCB.2003.811121
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, a method of classification referred to as the Bayesian data reduction algorithm (BDRA) is developed. The algorithm is based on the assumption that the discrete symbol probabilities of each class are a prion uniformly Dirichlet distributed, and it employs a "greedy" approach (which is similar to a backward sequential feature search) for reducing irrelevant features from the training data of each class. Notice that reducing irrelevant features is synonymous here with selecting those features that provide best classification performance; the metric for making data-reducing decisions is an analytic for the probability of error conditioned on the training data. To illustrate its performance, the BDRA is applied both to simulated and to real data, and it is also compared to other classification methods. Further, the algorithm is extended to deal with the problem of missing features in the data. Results demonstrate that the BDRA performs well despite its relative simplicity. This is significant because the BDRA differs from many other classifiers; as opposed to adjusting the model to obtain a "best fit" for the data, the data, through its quantization, is itself adjusted.
引用
收藏
页码:448 / 464
页数:17
相关论文
共 60 条
[1]  
ALKOOT FM, 2000, P INT C INF FUS JUL
[2]  
Bagenstoss PM, 1999, IEEE T SIGNAL PROCES, V47, P3428, DOI 10.1109/78.806092
[3]  
Bar-Shalom Y., 1995, MULTITARGET MULTISEN
[4]   Estimating the number of undetected errors: Bayesian model selection [J].
Basu, S ;
Ebrahimi, N .
NINTH INTERNATIONAL SYMPOSIUM ON SOFTWARE RELIABILITY ENGINEERING, PROCEEDINGS, 1998, :22-31
[5]   An empirical comparison of voting classification algorithms: Bagging, boosting, and variants [J].
Bauer, E ;
Kohavi, R .
MACHINE LEARNING, 1999, 36 (1-2) :105-139
[6]  
BAY SD, 1999, INTELL DATA ANAL, V3, P191, DOI DOI 10.1016/S1088-467X(99)00018-9
[7]  
Bellman R, 1961, ADAPTIVE CONTROL PRO, DOI DOI 10.1515/9781400874668
[8]  
Bernardo J. M., 1994, BAYESIAN THEORY
[9]  
Bishop C. M., 1995, NEURAL NETWORKS PATT
[10]  
BLUM AL, 1997, P ART INT, P245