Theory of Neural Nets Seminar: 14th June 2021

Event details
Date | 14.06.2021 |
Hour | 16:30 › 17:30 |
Speaker | Stephan Wojtowytsch (Princeton University) |
Location | Online |
Category | Conferences - Seminars |
This seminar consists of talks about current research on the theory of neural networks. Every session lasts one hour and comprises a talk (about 30 minutes) followed by a discussion with questions from the audience.
Speaker: Stephan Wojtowytsch (Princeton University)
Title: Stochastic gradient descent for noise with ML-type scaling
Abstract: In the literature on stochastic gradient descent, there are two types of convergence results: (1) SGD finds minimizers of convex objective functions and (2) SGD finds critical points of smooth objective functions. Classical results are obtained under the assumption that the stochastic noise is L^2-bounded and that the learning rate decays to zero at a suitable speed. We show that, if the objective landscape and noise possess certain properties which are reminiscent of deep learning problems, then we can obtain global convergence guarantees of first type under second type assumptions for a fixed (small, but positive) learning rate. The convergence is exponential, but with a large random coefficient. If the learning rate exceeds a certain threshold, we discuss minimum selection by studying the invariant distribution of a continuous time SGD model. We show that at a critical threshold, SGD prefers minimizers where the objective function is ‘flat’ in a precise sense.
Links
Practical information
- Expert
- Free
Contact
- François Ged: francois.ged[at]epfl.ch