Two timescale analysis of the Alopex algorithm for optimization

被引:12
作者
Sastry, PS [1 ]
Magesh, M
Unnikrishnan, KP
机构
[1] Indian Inst Sci, Dept Elect Engn, Bangalore 560012, Karnataka, India
[2] GM Corp, R&D Ctr, Warren, MI 48090 USA
关键词
D O I
10.1162/089976602760408044
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Alopex is a correlation-based gradient-free optimization technique useful in many learning problems. However, there are no analytical results on the asymptotic behavior of this algorithm. This article presents a new version of Alopex that can be analyzed using techniques of two timescale stochastic approximation method. It is shown that the algorithm asymptotically behaves like a gradient-descent method, though it does not need (or estimate) any gradient information. It is also shown, through simulations, that the algorithm is quite effective.
引用
收藏
页码:2729 / 2750
页数:22
相关论文
共 29 条
[1]  
[Anonymous], 2000, SUPERVISED UNSUPERVI
[2]  
BARTO AG, 1987, P IEEE 1 ANN C NEUR, P629
[3]  
BERTSEKAS DP, 1995, NONLINEAR PROGRAMMIN
[4]   Stochastic approximation algorithms: Overview and recent trends [J].
Bharath, B ;
Borkar, VS .
SADHANA-ACADEMY PROCEEDINGS IN ENGINEERING SCIENCES, 1999, 24 (4-5) :425-452
[5]  
Bia A, 2001, Int J Neural Syst, V11, P497, DOI 10.1016/S0129-0657(01)00092-8
[6]   Stochastic approximation with two time scales [J].
Borkar, VS .
SYSTEMS & CONTROL LETTERS, 1997, 29 (05) :291-294
[7]  
Brieman L., 1984, WADSWORTH INC
[8]  
BRODLEY CE, 1995, MACH LEARN, V19, P45, DOI 10.1007/BF00994660
[9]  
Dembo A, 1990, IEEE Trans Neural Netw, V1, P58, DOI 10.1109/72.80205
[10]  
DRAPER CS, 1951, PRINCIPLES OPTIMALIS