The interplay between data structure and neural networks: going beyond Gaussian models

Thumbnail

Event details

Date 27.01.2023
Hour 13:1514:15
Speaker Sebastian Goldt
Location
Category Conferences - Seminars
Event Language English

Neural networks are powerful feature extractors - but what do they actually learn from their data? We discuss two recent works on this question, with a focus on the importance of non-Gaussian statistics for neural networks. We first develop a simple model for images and show that a neural network trained on these images can learn a convolution from scratch. This pattern-formation process is driven by a combination of translation-invariance of the "images" and the non-Gaussian, higher-order statistics of the inputs. Second, we conjecture a "distributional simplicity bias" whereby neural networks learn increasingly complex distributions of their inputs during training. We present analytical and experimental evidence for this conjecture, going from a simple perceptron up to deep ResNets and visual transformers.

Practical information

  • Informed public
  • Free

Contact

Share