A tutorial on stochastic approximation algorithms for training Restricted Boltzmann Machines and Deep Belief Nets

Information Theory and Applications Workshop(2010)

引用 73|浏览28
暂无评分
摘要
In this study, we provide a direct comparison of the Stochastic Maximum Likelihood algorithm and Contrastive Divergence for training Restricted Boltzmann Machines using the MNIST data set. We demonstrate that Stochastic Maximum Likelihood is superior when using the Restricted Boltzmann Machine as a classifier, and that the algorithm can be greatly improved using the technique of iterate averaging from the field of stochastic approximation. We further show that training with optimal parameters for classification does not necessarily lead to optimal results when Restricted Boltzmann Machines are stacked to form a Deep Belief Network. In our experiments we observe that fine tuning a Deep Belief Network significantly changes the distribution of the latent data, even though the parameter changes are negligible.
更多
查看译文
关键词
approximation theory,belief networks,iterative methods,maximum likelihood estimation,stochastic processes,contrastive divergence,deep belief network,iterate averaging,maximum likelihood algorithm,restricted Boltzmann machines,stochastic approximation
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要