Spectral pruning of fully connected layers.
Journal
Scientific reports
ISSN: 2045-2322
Titre abrégé: Sci Rep
Pays: England
ID NLM: 101563288
Informations de publication
Date de publication:
01 07 2022
01 07 2022
Historique:
received:
25
02
2022
accepted:
13
06
2022
entrez:
1
7
2022
pubmed:
2
7
2022
medline:
7
7
2022
Statut:
epublish
Résumé
Training of neural networks can be reformulated in spectral space, by allowing eigenvalues and eigenvectors of the network to act as target of the optimization instead of the individual weights. Working in this setting, we show that the eigenvalues can be used to rank the nodes' importance within the ensemble. Indeed, we will prove that sorting the nodes based on their associated eigenvalues, enables effective pre- and post-processing pruning strategies to yield massively compacted networks (in terms of the number of composing neurons) with virtually unchanged performance. The proposed methods are tested for different architectures, with just a single or multiple hidden layers, and against distinct classification tasks of general interest.
Identifiants
pubmed: 35778586
doi: 10.1038/s41598-022-14805-7
pii: 10.1038/s41598-022-14805-7
pmc: PMC9249877
doi:
Types de publication
Journal Article
Langues
eng
Sous-ensembles de citation
IM
Pagination
11201Informations de copyright
© 2022. The Author(s).
Références
Nat Neurosci. 2004 Jul;7(7):691-2
pubmed: 15220926
Nat Commun. 2021 Feb 26;12(1):1330
pubmed: 33637729
Phys Rev E. 2021 Nov;104(5-1):054312
pubmed: 34942751