Affiliation:
1. Department of Electrical and Computer Engineering, Utah State University, Logan, UT 84322, USA
Abstract
Adaptive natural gradient learning avoids singularities in the parameter
space of multilayer perceptrons. However, it requires a larger number
of additional parameters than ordinary backpropagation in the form of
the Fisher information matrix. This paper describes a new approach to
natural gradient learning that uses a smaller Fisher information matrix.
It also uses a prior distribution on the neural network parameters and an
annealed learning rate. While this new approach is computationally simpler,
its performance is comparable to that of adaptive natural gradient
learning.
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献