Dynamics of Learning with Restricted Training Sets I: General Theory

Physics – Condensed Matter – Disordered Systems and Neural Networks

Scientific paper

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

39 pages, LaTeX

Scientific paper

10.1103/PhysRevE.62.5444

We study the dynamics of supervised learning in layered neural networks, in the regime where the size $p$ of the training set is proportional to the number $N$ of inputs. Here the local fields are no longer described by Gaussian probability distributions and the learning dynamics is of a spin-glass nature, with the composition of the training set playing the role of quenched disorder. We show how dynamical replica theory can be used to predict the evolution of macroscopic observables, including the two relevant performance measures (training error and generalization error), incorporating the old formalism developed for complete training sets in the limit $\alpha=p/N\to\infty$ as a special case. For simplicity we restrict ourselves in this paper to single-layer networks and realizable tasks.

No associations

LandOfFree

Say what you really think

Search LandOfFree.com for scientists and scientific papers. Rate them and share your experience with other people.

Rating

Dynamics of Learning with Restricted Training Sets I: General Theory does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.

If you have personal experience with Dynamics of Learning with Restricted Training Sets I: General Theory, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Dynamics of Learning with Restricted Training Sets I: General Theory will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFWR-SCP-O-423660

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.