Finite-size effects in on-line learning of multilayer neural networks

David Barber, David Saad, Peter Sollich

Research output: Contribution to journalArticlepeer-review

Abstract

We complement recent advances in thermodynamic limit analyses of mean on-line gradient descent learning dynamics in multi-layer networks by calculating fluctuations possessed by finite dimensional systems. Fluctuations from the mean dynamics are largest at the onset of specialisation as student hidden unit weight vectors begin to imitate specific teacher vectors, increasing with the degree of symmetry of the initial conditions. In light of this, we include a term to stimulate asymmetry in the learning process, which typically also leads to a significant decrease in training time.
Original languageEnglish
Pages (from-to)151-156
Number of pages6
JournalEurophysics Letters
Volume34
Issue number2
Publication statusPublished - Apr 1996

Bibliographical note

Copyright of EDP Sciences

Keywords

  • probability theory
  • stochastic processes
  • and statistics

Fingerprint

Dive into the research topics of 'Finite-size effects in on-line learning of multilayer neural networks'. Together they form a unique fingerprint.

Cite this