Reinforcement Interval Type-2 Fuzzy Controller Design by Online Rule Generation and Q-Value-Aided Ant Colony Optimization

被引:65
作者
Juang, Chia-Feng [1 ]
Hsu, Chia-Hung [1 ]
机构
[1] Natl Chung Hsing Univ, Dept Elect Engn, Taichung 402, Taiwan
来源
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS | 2009年 / 39卷 / 06期
关键词
Ant colony optimization (ACO); fuzzy Q-learning; interval type-2 fuzzy sets; reinforcement learning; type-2 fuzzy systems; NEURAL-NETWORK; SYMBIOTIC EVOLUTION; INFERENCE NETWORK; SYSTEM; INTERPRETABILITY; ALGORITHM;
D O I
10.1109/TSMCB.2009.2020569
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper proposes a new reinforcement-learning method using online rule generation and Q-value-aided ant colony optimization (ORGQACO) for fuzzy controller design. The fuzzy controller is based on an interval type-2 fuzzy system (IT2FS). The antecedent part in the designed IT2FS uses interval type-2 fuzzy sets to improve controller robustness to noise. There are initially no fuzzy rules in the IT2FS. The ORGQACO concurrently designs both the structure and parameters of an IT2FS. We propose an online interval type-2 rule generation method for the evolution of system structure and flexible partitioning of the input space. Consequent part parameters in an IT2FS are designed using Q-values and the reinforcement local-global ant colony optimization algorithm. This algorithm selects the consequent part from a set of candidate actions according to ant pheromone trails and Q-values, both of which are updated using reinforcement signals. The ORGQACO design method is applied to the following three control problems: 1) truck-backing control; 2) magnetic-levitation control; and 3) chaotic-system control. The ORGQACO is compared with other reinforcement-learning methods to verify its efficiency and effectiveness. Comparisons with type-1 fuzzy systems verify the noise robustness property of using an IT2FS.
引用
收藏
页码:1528 / 1542
页数:15
相关论文
共 44 条
[1]   Hybrid learning models to get the interpretability-accuracy trade-off in fuzzy modeling [J].
Alcalá, R ;
Alcalá-Fdez, J ;
Casillas, J ;
Cordón, O ;
Herrera, F .
SOFT COMPUTING, 2006, 10 (09) :717-734
[2]  
[Anonymous], P IJCNN 1990
[3]  
Astudillo L, 2007, LECT NOTES COMPUT SC, V4529, P594
[4]   LEARNING AND TUNING FUZZY-LOGIC CONTROLLERS THROUGH REINFORCEMENTS [J].
BERENJI, HR ;
KHEDKAR, P .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1992, 3 (05) :724-740
[5]   The hyper-cube framework for ant colony optimization [J].
Blum, C ;
Dorigo, M .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2004, 34 (02) :1161-1172
[6]   Learning cooperative linguistic fuzzy rules using the best-worst ant system algorithm [J].
Casillas, J ;
Cordón, O ;
de Viana, IF ;
Herrera, F .
INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2005, 20 (04) :433-452
[7]   A self-learning fuzzy logic controller using genetic algorithms with reinforcements [J].
Chiang, CK ;
Chung, HY ;
Lin, JJ .
IEEE TRANSACTIONS ON FUZZY SYSTEMS, 1997, 5 (03) :460-467
[8]   Support vector learning mechanism for fuzzy rule-based modeling: A new approach [J].
Chiang, JH ;
Hao, PY .
IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2004, 12 (01) :1-12
[9]   Semantic constraints for membership function optimization [J].
de Oliveira, JV .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART A-SYSTEMS AND HUMANS, 1999, 29 (01) :128-138
[10]  
Dorigo M, 2004, ANT COLONY OPTIMIZATION, P1