Identifying the impact of decision variables for nonlinear classification tasks

被引:13
作者
Kim, SH [1 ]
Shin, SW
机构
[1] Korea Adv Inst Sci & Technol, Grad Sch Management, Seoul, South Korea
[2] Samsung SDS Co Ltd, Seoul, South Korea
关键词
feature weighting; similarity assessment; k-nearest neighbor; lazy learning; artificial neural network; genetic algorithms;
D O I
10.1016/S0957-4174(99)00062-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents a novel procedure to improve a class of learning systems known as lazy learning algorithms by optimizing the selection of variables and their attendant weights through an artificial neural network and a genetic algorithm. The procedure utilizes its previous knowledge base-also called a case base-to select an effective subset for adaptation. In particular, the procedure explores a space of N variables and generates a reduced space of M dimensions. This is achieved through clustering and compaction. The clustering stage involves the minimization of distances among individuals within the same class while maximizing the distances among different classes. The compaction stage involves the elimination of the irrelevant or redundant feature dimensions. To achieve these two goals concurrently through the evolutionary process, new measures of fitness have been developed. The metrics lead to procedures which exhibit superior characteristics in terms of both accuracy and efficiency. The efficiency springs from a reduction in the number of features required for analysis, thereby saving on computational cost as well as data collection requirements. The utility of the new techniques is validated against a variety of data sets from natural and commercial sources. (C) 2000 Published by Elsevier Science Ltd. All rights reserved.
引用
收藏
页码:201 / 214
页数:14
相关论文
共 35 条
[1]   DATABASE MINING - A PERFORMANCE PERSPECTIVE [J].
AGRAWAL, R ;
IMIELINSKI, T ;
SWAMI, A .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 1993, 5 (06) :914-925
[2]   INSTANCE-BASED LEARNING ALGORITHMS [J].
AHA, DW ;
KIBLER, D ;
ALBERT, MK .
MACHINE LEARNING, 1991, 6 (01) :37-66
[3]  
[Anonymous], 1989, GENETIC ALGORITHM SE
[4]  
[Anonymous], 1997, APPL CASE BASED REAS
[5]  
[Anonymous], PATTER CLASSIFICATIO
[6]  
[Anonymous], 1994, MACH LEARN P 1994
[7]  
[Anonymous], 1995, PRACTICAL HDB GENETI
[8]  
[Anonymous], IJCAI
[9]   DETERMINING INPUT FEATURES FOR MULTILAYER PERCEPTRONS [J].
BELUE, LM ;
BAUER, KW .
NEUROCOMPUTING, 1995, 7 (02) :111-121
[10]   FAST GENETIC SELECTION OF FEATURES FOR NEURAL NETWORK CLASSIFIERS [J].
BRILL, FZ ;
BROWN, DE ;
MARTIN, WN .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1992, 3 (02) :324-328