PS : Umut Simsekli (INRIA et ENS Paris) : Implicit Compressibility of Overparametrized Neural Networks Trained with Heavy-Tailed SGD

Chargement Évènements

« Tous les Évènements

  • Cet évènement est passé.

PS : Umut Simsekli (INRIA et ENS Paris) : Implicit Compressibility of Overparametrized Neural Networks Trained with Heavy-Tailed SGD

14 février / 11:30 - 12:30

Neural network compression has been an increasingly important subject, not only due to its practical relevance, but also due to its theoretical implications, as there is an explicit connection between compressibility and generalization error. In this talk, I will present a simple modification for SGD, such that the outputs of the algorithm will be provably compressible without making any nontrivial assumptions. We will consider a one-hidden-layer neural network trained with SGD, and show that if we inject additive heavy-tailed noise to the iterates at each iteration, for any compression rate, there exists a level of overparametrization such that the output of the algorithm will be compressible with high probability. To achieve this result, we make two main technical contributions: (i) we prove a “propagation of chaos” result for a class of heavy-tailed stochastic differential equations, and (ii) we derive error estimates for their Euler discretization. Our experiments suggest that the proposed approach not only achieves increased compressibility with various models and datasets, but also leads to robust test performance under pruning, even in more realistic architectures that lie beyond our theoretical setting. The talk is based on the following article: https://arxiv.org/pdf/2306.08125.pdf

PS : Umut Simsekli (INRIA et ENS Paris) : Implicit Compressibility of Overparametrized Neural Networks Trained with Heavy-Tailed SGD

Détails

Date :
14 février
Heure :
11:30 - 12:30
Catégorie d’Évènement:

Lieu

Bâtiment Fermat, salle 4205

Organisateurs

Ester Mariucci
Emmanuel Rio