Label noise (stochastic) gradient descent implicitly solves the Lasso for quadratic parametrisation - École des Ponts ParisTech Accéder directement au contenu
Communication Dans Un Congrès Année : 2022

Label noise (stochastic) gradient descent implicitly solves the Lasso for quadratic parametrisation

Résumé

Understanding the implicit bias of training algorithms is of crucial importance in order to explain the success of overparametrised neural networks. In this paper, we study the role of the label noise in the training dynamics of a quadratically parametrised model through its continuous time version. We explicitly characterise the solution chosen by the stochastic flow and prove that it implicitly solves a Lasso program. To fully complete our analysis, we provide nonasymptotic convergence guarantees for the dynamics as well as conditions for support recovery. We also give experimental results which support our theoretical claims. Our findings highlight the fact that structured noise can induce better generalisation and help explain the greater performances of stochastic dynamics as observed in practice.

Dates et versions

hal-03701409 , version 1 (22-06-2022)

Identifiants

Citer

Loucas Pillaud-Vivien, Julien Reygner, Nicolas Flammarion. Label noise (stochastic) gradient descent implicitly solves the Lasso for quadratic parametrisation. Thirty Fifth Conference on Learning Theory, Jul 2022, Londres, United Kingdom. pp.2127-2159. ⟨hal-03701409⟩
16 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More