Chapter 17 Geometry of neural networks: Natural gradient for learning

作者: K. Fukumizu

DOI: 10.1016/S1383-8121(01)80020-8

关键词:

摘要: Publisher Summary This chapter discusses the geometric structure of multilayer networks and dynamics learning caused by their structure. A neural network has several layers, including input output layer, connections between neighboring layers. It realizes a function from vector to vector—for example, perceptrons radial basis functions. The explains that statistical formulation enables introduction Riemannian metric on model. is essentially an intrinsic quantity manifold, which defined in space all probabilities. natural gradient direction gives steepest descent for change small, fixed length measured metric. In introducing gradient, emphasizes viewpoint parameter not necessarily Euclidean but more can be introduced many problems. case parametric estimation, Fisher information this.

参考文章(25)
Klaus Jänich, Theodor Bröcker, Introduction to differential topology Cambridge University Press. ,(1982)
Christopher M. Bishop, Neural networks for pattern recognition ,(1995)
K. Hagiwara, N. Toda, S. Usui, On the problem of applying AIC to determine the structure of a layered feedforward neural network international joint conference on neural network. ,vol. 3, pp. 2263- 2266 ,(1993) , 10.1109/IJCNN.1993.714176
Shun-ichi Amari, Natural gradient works efficiently in learning Neural Computation. ,vol. 10, pp. 177- 202 ,(1998) , 10.1162/089976698300017746
K. Fukumizu, S. Amari, Local minima and plateaus in hierarchical structures of multilayer perceptions Neural Networks. ,vol. 13, pp. 317- 327 ,(2000) , 10.1016/S0893-6080(00)00009-5
David Saad, Sara A. Solla, On-line learning in soft committee machines Physical Review E. ,vol. 52, pp. 4225- 4243 ,(1995) , 10.1103/PHYSREVE.52.4225