The interplay between data structure and neural networks: going beyond Gaussian models

Event details
Date | 27.01.2023 |
Hour | 13:15 › 14:15 |
Speaker | Sebastian Goldt |
Location | |
Category | Conferences - Seminars |
Event Language | English |
Neural networks are powerful feature extractors - but what do they actually learn from their data? We discuss two recent works on this question, with a focus on the importance of non-Gaussian statistics for neural networks. We first develop a simple model for images and show that a neural network trained on these images can learn a convolution from scratch. This pattern-formation process is driven by a combination of translation-invariance of the "images" and the non-Gaussian, higher-order statistics of the inputs. Second, we conjecture a "distributional simplicity bias" whereby neural networks learn increasingly complex distributions of their inputs during training. We present analytical and experimental evidence for this conjecture, going from a simple perceptron up to deep ResNets and visual transformers.
Practical information
- Informed public
- Free
Contact
- Lénaïc Chizat: [email protected] François Ged: [email protected]