High-dimensional dynamics of generalization error in neural networks: implications for experience replay
- 👤 Speaker: Dr. Andrew Saxe, University of Oxford
- 📅 Date & Time: Friday 01 February 2019, 12:30 - 13:30
- 📍 Venue: Cambridge University Engineering Department, CBL, BE4-38 (http://learning.eng.cam.ac.uk/Public/Directions)
Abstract
Learning even a simple task can engage huge numbers of neurons across the cortical hierarchy. How do neuronal networks manage to generalize from a small number of examples, despite having large numbers of tunable synapses? And how does depth—the serial propagation of signals through layered structure—impact a learning system? I will describe results emerging from the analysis of deep linear neural networks. Deep linear networks are a simple model class that retain many features of the full nonlinear setting, including a nonconvex error surface and nonlinear learning trajectories. In this talk I will focus on their generalization error, using random matrix theory to analyze the cognitively-relevant “high-dimensional” regime, where the number of training examples is on the order of or even less than the number of adjustable synapses. Consistent with the striking performance of very large deep network models in practice, I show that good generalization is possible in overcomplete networks due to implicit regularization in the dynamics of gradient descent. Overtraining is worst at intermediate network sizes, when the effective number of free parameters equals the number of samples, and can be reduced by making a network smaller or larger. I identify two novel phenomena underlying this behavior in linear networks: first, there is a frozen subspace of the weights in which no learning occurs under gradient descent; and second, the statistical properties of the high-dimensional regime yield better-conditioned input correlations which protect against overtraining. Turning to the impact of depth, the theory reveals a trade-off between training speed and generalization performance in deep neural networks, and I confirm this speed-accuracy trade-off through simulations. Finally, I will describe an application of these results to experience replay during sleep. The consolidation of learning during sleep is thought to arise from the replay of stored experiences between hippocampus and neocortex. Why is this complex strategy beneficial? As a simple model of this process, we compare the dynamics arising from online learning, in which each example is used once and discarded; and batch learning, in which all examples are stored (for instance, in hippocampus) and replayed repeatedly (for instance, during sleep). While these two strategies yield similar performance when training experience is abundant, we find that replay can be decisively better when training experience is scarce. Our results suggest a normative explanation for a two-stage memory system: replay can enable better generalization from limited training experience.
Series This talk is part of the Computational Neuroscience series.
Included in Lists
- All Talks (aka the CURE list)
- Biology
- Biology
- bld31
- Cambridge Centre for Data-Driven Discovery (C2D3)
- Cambridge Forum of Science and Humanities
- Cambridge Language Sciences
- Cambridge Neuroscience Seminars
- CamBridgeSens
- Cambridge talks
- Cambridge University Engineering Department, CBL, BE4-38 (http://learning.eng.cam.ac.uk/Public/Directions)
- CBL important
- Chris Davis' list
- Computational and Biological Learning Seminar Series
- Computational Neuroscience
- Creating transparent intact animal organs for high-resolution 3D deep-tissue imaging
- custom
- dh539
- dh539
- Featured lists
- Guy Emerson's list
- Hanchen DaDaDash
- Inference Group Journal Clubs
- Inference Group Summary
- Information Engineering Division seminar list
- Interested Talks
- Joint Machine Learning Seminars
- Life Science
- Life Science Interface Seminars
- Life Sciences
- Life Sciences
- Machine Learning @ CUED
- Machine Learning Summary
- ME Seminar
- ML
- my_list
- ndk22's list
- Neuroscience
- Neuroscience Seminars
- Neuroscience Seminars
- ob366-ai4er
- other talks
- Quantum Matter Journal Club
- Required lists for MLG
- rp587
- se456's list
- Seminar
- Simon Baker's List
- Stem Cells & Regenerative Medicine
- TQS Journal Clubs
- Trust & Technology Initiative - interesting events
- yk373's list
- yk449
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)

Dr. Andrew Saxe, University of Oxford
Friday 01 February 2019, 12:30-13:30