Quickly Generating Representative Samples from an RBM-Derived Process

被引:39
作者
Breuleux, Olivier [1 ]
Bengio, Yoshua [1 ]
Vincent, Pascal [1 ]
机构
[1] Univ Montreal, Dept Informat & Rech Operat, Montreal, PQ H3T 1J8, Canada
关键词
D O I
10.1162/NECO_a_00158
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Two recently proposed learning algorithms, herding and fast persistent contrastive divergence (FPCD), share the following interesting characteristic: they exploit changes in the model parameters while sampling in order to escape modes and mix better during the sampling process that is part of the learning algorithm. We justify such approaches as ways to escape modes while keeping approximately the same asymptotic distribution of the Markov chain. In that spirit, we extend FPCD using an idea borrowed from Herding in order to obtain a pure sampling algorithm, which we call the rates-FPCD sampler. Interestingly, this sampler can improve the model as we collect more samples, since it optimizes a lower bound on the log likelihood of the training data. We provide empirical evidence that this new algorithm displays substantially better and more robust mixing than Gibbs sampling.
引用
收藏
页码:2058 / 2073
页数:16
相关论文
共 13 条
[1]  
[Anonymous], 2010, P 13 INT C ARTIFICIA
[2]   Justifying and Generalizing Contrastive Divergence [J].
Bengio, Yoshua ;
Delalleau, Olivier .
NEURAL COMPUTATION, 2009, 21 (06) :1601-1621
[3]  
Borkar VS., 2009, Stochastic Approximation: A Dynamical Systems Viewpoint
[4]  
Breuleux O., 2010, 1349 DIRO U MONTR
[5]  
Hinton G. E., 1984, TRCMUCS84119
[6]   Training products of experts by minimizing contrastive divergence [J].
Hinton, GE .
NEURAL COMPUTATION, 2002, 14 (08) :1771-1800
[7]  
Hinton GE, 1999, IEE CONF PUBL, P1, DOI 10.1049/cp:19991075
[8]   A fast learning algorithm for deep belief nets [J].
Hinton, Geoffrey E. ;
Osindero, Simon ;
Teh, Yee-Whye .
NEURAL COMPUTATION, 2006, 18 (07) :1527-1554
[9]  
Tieleman, 2008, P 25 INT C MACH LEAR, P1064, DOI DOI 10.1145/1390156.1390290
[10]  
Tieleman T, 2009, P 26 ANN INT C MACH, P1033