Entropy and mutual information in models of deep neural networks

Thumbnail

Event details

Date 02.11.2018
Hour 14:0015:00
Speaker Marylou Gabrié
Location
Category Conferences - Seminars

The successes and the multitude of applications of deep learning methods have spurred efforts towards quantitative modeling of the performance of deep neural networks. In particular, an information-theoretic approach linking generalization capabilities to compression has been receiving increasing interest. Nevertheless, it is in practice computationally intractable to compute entropies and mutual informations in industry-sized neural networks. In this talk, we will consider instead a class of models of deep neural networks, for which an expression for these information-theoretic quantities can be derived from the replica method. We will examine how mutual informations between hidden and input variables can be reported along the training of such neural networks on synthetic datasets. Finally we will discuss the numerical results of a few training experiments.

This work was done in collaboration with Andre Manoel (Owkin), Clément Luneau (EPFL), Jean Barbier (EPFL), Nicolas Macris (EPFL), Florent Krzakala (LPS ENS) and Lenka Zdeborova (IPHT CEA).

Practical information

  • General public
  • Free

Organizer

  • Levent Sagun

Contact

  • Corinne Weibel

Event broadcasted in

Share