Abstract
We analyse natural gradient learning in a two-layer feed-forward neural network using a statistical mechanics framework which is appropriate for large input dimension. We find significant improvement over standard gradient descent in both the transient and asymptotic phases of learning.
Original language | English |
---|---|
Title of host publication | Proceedings of the 8th International Conference on Artificial Neural Networks |
Editors | L. Niklasson, M. Boden, T. Ziemke |
Publisher | Springer |
Pages | 165-170 |
Number of pages | 6 |
Volume | 1 |
ISBN (Print) | 3540762639 |
DOIs | |
Publication status | Published - 1 Sept 1998 |
Bibliographical note
The original publication is available at www.springerlink.comKeywords
- natural gradient
- statistical mechanics
- gradient descent
- transient
- asymptotic