288 pp. per issue
6 x 9, illustrated
2014 Impact factor:

Neural Computation

September 2008, Vol. 20, No. 9, Pages 2238-2252
(doi: 10.1162/neco.2008.01-07-430)
© 2008 Massachusetts Institute of Technology
Contrastive Divergence in Gaussian Diffusions
Article PDF (101.08 KB)

This letter presents an analysis of the contrastive divergence (CD) learning algorithm when applied to continuous-time linear stochastic neural networks. For this case, powerful techniques exist that allow a detailed analysis of the behavior of CD. The analysis shows that CD converges to maximum likelihood solutions only when the network structure is such that it can match the first moments of the desired distribution. Otherwise, CD can converge to solutions arbitrarily different from the log-likelihood solutions, or they can even diverge. This result suggests the need to improve our theoretical understanding of the conditions under which CD is expected to be well behaved and the conditions under which it may fail. In, addition the results point to practical ideas on how to improve the performance of CD.