Representational power of restricted Boltzmann machines and deep belief networks

被引:860
作者
Le Roux, Nicolas [1 ]
Bengio, Yoshua [1 ]
机构
[1] Univ Montreal, Dept Informat & Rech Operat, Montreal, PQ H3C 3J7, Canada
关键词
D O I
10.1162/neco.2008.04-07-510
中图分类号
TP18 [人工智能理论];
学科分类号
081104 [模式识别与智能系统]; 0812 [计算机科学与技术]; 0835 [软件工程]; 1405 [智能科学与技术];
摘要
Deep belief networks (DBN) are generative neural network models with many layers of hidden explanatory factors, recently introduced by Hinton, Osindero, and Teh (2006) along with a greedy layer-wise unsupervised learning algorithm. The building block of a DBN is a probabilistic model called a restricted Boltzmann machine (RBM), used to represent one layer of the model. Restricted Boltzmann machines are interesting because inference is easy in them and because they have been successfully used as building blocks for training deeper models. We first prove that adding hidden units yields strictly improved modeling power, while a second theorem shows that RBMs are universal approximators of discrete distributions. We then study the question of whether DBNs with more layers are strictly more powerful in terms of representational power. This suggests a new and less greedy criterion for training RBMs Within DBNs.
引用
收藏
页码:1631 / 1649
页数:19
相关论文
共 18 条
[1]
ACKLEY DH, 1985, COGNITIVE SCI, V9, P147
[2]
SIGMA-11-FORMULAE ON FINITE STRUCTURES [J].
AJTAI, M .
ANNALS OF PURE AND APPLIED LOGIC, 1983, 24 (01) :1-48
[3]
[Anonymous], 2005, AISTATS BRIDGETOWN B
[4]
[Anonymous], 2007, ADV NEURAL INFORM PR
[5]
[Anonymous], 1996, SOC ENV TOXICOLOGY C
[6]
Bengio Y., 2006, Advances in Neural Information Processing Systems, V18, P107
[7]
Bengio Y., 2007, LARGE SCALE KERNEL M
[8]
Bengio Yoshua, 2006, Advances in Neural Information Processing Systems 18 NIPS'05, P115
[9]
Hastad Johan, 1987, Computational limitations for small-depth circuits
[10]
Reducing the dimensionality of data with neural networks [J].
Hinton, G. E. ;
Salakhutdinov, R. R. .
SCIENCE, 2006, 313 (5786) :504-507