Abstract
The Langevin updating rule, in which noise is added to the weights during learning, is presented and shown to improve learning on problems with initially ill-conditioned Hessians. This is particularly important for multilayer perceptrons with many hidden layers, that often have ill-conditioned Hessians. In addition, Manhattan updating is shown to have a similar effect.
This content is only available as a PDF.
© 1994 Massachusetts Institute of Technology
1994
You do not currently have access to this content.