Batch Normalization Provably Avoids Rank Collapse for Randomly Initialised Deep Networks - INRIA - Institut National de Recherche en Informatique et en Automatique Accéder directement au contenu
Communication Dans Un Congrès Année : 2020

Batch Normalization Provably Avoids Rank Collapse for Randomly Initialised Deep Networks

Résumé

Randomly initialized neural networks are known to become harder to train with increasing depth, unless architectural enhancements like residual connections and batch normalization are used. We here investigate this phenomenon by revisiting the connection between random initialization in deep networks and spectral instabilities in products of random matrices. Given the rich literature on random matrices, it is not surprising to find that the rank of the intermediate representations in unnormalized networks collapses quickly with depth. In this work we highlight the fact that batch normalization is an effective strategy to avoid rank collapse for both linear and ReLU networks. Leveraging tools from Markov chain theory, we derive a meaningful lower rank bound in deep linear networks. Empirically, we also demonstrate that this rank robustness generalizes to ReLU nets. Finally, we conduct an extensive set of experiments on real-world data sets, which confirm that rank stability is indeed a crucial condition for training modern-day deep neural architectures.
Fichier principal
Vignette du fichier
NeurIPS-2020-batch-normalization-provably-avoids-ranks-collapse-for-randomly-initialised-deep-networks-Paper (4).pdf (742.08 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03454386 , version 1 (29-11-2021)

Identifiants

  • HAL Id : hal-03454386 , version 1

Citer

Hadi Daneshmand, Jonas Kohler, Francis Bach, Thomas Hofmann, Aurelien Lucchi. Batch Normalization Provably Avoids Rank Collapse for Randomly Initialised Deep Networks. NeurIPS 2020 - Thirty-fourth Conference on Neural Information Processing Systems, Dec 2020, Virtual, France. ⟨hal-03454386⟩
45 Consultations
37 Téléchargements

Partager

Gmail Facebook X LinkedIn More