Stochastic approximation driven particle swarm optimization with simultaneous perturbation - Who will guide the guide?

被引:3
作者
Kiranyaz, Serkan [1 ]
Ince, Turker [2 ]
Gabbouj, Moncef [1 ]
机构
[1] Tampere Univ Technol, Dept Signal Proc, FIN-33101 Tampere, Finland
[2] Izmir Univ Econ, Dept Elect & Telecommun Engn, TR-35330 Izmir, Turkey
关键词
Particle swarm optimization; Stochastic approximation; Multi-dimensional search; Gradient descent; GLOBAL OPTIMIZATION; ALGORITHMS;
D O I
10.1016/j.asoc.2010.07.022
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The need for solving multi-modal optimization problems in high dimensions is pervasive in many practical applications. Particle swarm optimization (PSO) is attracting an ever-growing attention and more than ever it has found many application areas for many challenging optimization problems. It is, however, a known fact that PSO has a severe drawback in the update of its global best (gbest) particle, which has a crucial role of guiding the rest of the swarm. In this paper, we propose two efficient solutions to remedy this problem using a stochastic approximation (SA) technique. In the first approach, gbest is updated (moved) with respect to a global estimation of the gradient of the underlying (error) surface or function and hence can avoid getting trapped into a local optimum. The second approach is based on the formation of an alternative or artificial global best particle, the so-called aGB, which can replace the native gbest particle for a better guidance, the decision of which is held by a fair competition between the two. For this purpose we use simultaneous perturbation stochastic approximation (SPSA) for its low cost. Since SPSA is applied only to the gbest (not to the entire swarm), both approaches result thus in a negligible overhead cost for the entire PSO process. Both approaches are shown to significantly improve the performance of PSO over a wide range of non-linear functions, especially if SPSA parameters are well selected to fit the problem at hand. A major finding of the paper is that even if the SPSA parameters are not tuned well, results of SA-driven (SAD) PSO are still better than the best of PSO and SPSA. Since the problem of poor gbest update persists in the recently proposed extension of PSO, called multi-dimensional PSO (MD-PSO), both approaches are also integrated into MD-PSO and tested over a set of unsupervised data clustering applications. As in the basic PSO application, experimental results show that the proposed approaches significantly improved the quality of the MD-PSO clustering as measured by a validity index function. Furthermore, the proposed approaches are generic as they can be used with other PSO variants and applicable to a wide range of problems. (C) 2010 Elsevier B.V. All rights reserved.
引用
收藏
页码:2334 / 2347
页数:14
相关论文
共 42 条
[1]  
Abraham A., 2007, SOFT COMPUTING KNOWL, P279
[2]  
Angeline P. J., 1998, Evolutionary Programming VII. 7th International Conference, EP98. Proceedings, P601, DOI 10.1007/BFb0040811
[3]   Using selection to improve particle swarm optimization [J].
Angeline, PJ .
1998 IEEE INTERNATIONAL CONFERENCE ON EVOLUTIONARY COMPUTATION - PROCEEDINGS, 1998, :84-89
[4]  
[Anonymous], P IEEE C EV COMP
[5]  
[Anonymous], 1996, ADV KNOWLEDGE DISCOV
[6]  
[Anonymous], IEEE T EVOLUTIONARY
[7]  
[Anonymous], VESTERSTROM DIVERSIT
[8]  
[Anonymous], 2003, Genetic programming IV: routine human-competitive machine intelligence
[9]  
[Anonymous], 1995, 1995 IEEE INT C
[10]  
BACK T, 1995, FUZZY LOGIC SOFT COM, P3