Training restricted boltzmann machine using gradient fixing based algorithm

Fei Li, Xiaoguang Gao, Kaifang Wan

Research output: Contribution to journalArticlepeer-review

7 Scopus citations

Abstract

Most of the algorithms for training restricted Boltzmann machines (RBM) are based on Gibbs sampling. When the sampling algorithm is used to calculate the gradient, the sampling gradient is the approximate value of the true gradient and there is a big error between the sampling gradient and the true gradient, which seriously affects the training effect of the network. Aiming at this problem, this paper analysed the numerical error and orientation error between the approximate gradient and the true gradient. Their influence on the performance of network training is given then. An gradient fixing model was established. It was designed to adjust the numerical value and orientation of the approximate gradient and reduce the error. We also designed gradient fixing based Gibbs sampling training algorithm (GFGS) and gradient fixing based parallel tempering algorithm (GFPT), and the comparison experiment of the novel algorithms and the existing algorithms is given. It has been demonstrated that the new algorithms can effectively tackle the issue of gradient error, and can achieve higher training accuracy at a reasonable expense of computational runtime.

Original languageEnglish
Pages (from-to)694-703
Number of pages10
JournalChinese Journal of Electronics
Volume27
Issue number4
DOIs
StatePublished - 10 Jul 2018

Keywords

  • Deep Learning
  • Gibbs sampling training algorithm (GFGS)
  • Gradient fixing
  • Gradient fixing based parallel tempering algorithm (GFPT)
  • Restricted Boltzmann machine (RBM)

Fingerprint

Dive into the research topics of 'Training restricted boltzmann machine using gradient fixing based algorithm'. Together they form a unique fingerprint.

Cite this