Zero time waste in pre-trained early exit neural networks.
Conditional computation
Deep learning
Dynamic neural networks
Early-exiting networks
Zero waste models
Journal
Neural networks : the official journal of the International Neural Network Society
ISSN: 1879-2782
Titre abrégé: Neural Netw
Pays: United States
ID NLM: 8805018
Informations de publication
Date de publication:
Nov 2023
Nov 2023
Historique:
received:
22
05
2023
revised:
29
08
2023
accepted:
04
10
2023
medline:
13
11
2023
pubmed:
15
10
2023
entrez:
14
10
2023
Statut:
ppublish
Résumé
The problem of reducing processing time of large deep learning models is a fundamental challenge in many real-world applications. Early exit methods strive towards this goal by attaching additional Internal Classifiers (ICs) to intermediate layers of a neural network. ICs can quickly return predictions for easy examples and, as a result, reduce the average inference time of the whole model. However, if a particular IC does not decide to return an answer early, its predictions are discarded, with its computations effectively being wasted. To solve this issue, we introduce Zero Time Waste (ZTW), a novel approach in which each IC reuses predictions returned by its predecessors by (1) adding direct connections between ICs and (2) combining previous outputs in an ensemble-like manner. We conduct extensive experiments across various multiple modes, datasets, and architectures to demonstrate that ZTW achieves a significantly better accuracy vs. inference time trade-off than other early exit methods. On the ImageNet dataset, it obtains superior results over the best baseline method in 11 out of 16 cases, reaching up to 5 percentage points of improvement on low computational budgets.
Identifiants
pubmed: 37837747
pii: S0893-6080(23)00555-5
doi: 10.1016/j.neunet.2023.10.003
pii:
doi:
Types de publication
Journal Article
Langues
eng
Sous-ensembles de citation
IM
Pagination
580-601Informations de copyright
Copyright © 2023 The Author(s). Published by Elsevier Ltd.. All rights reserved.
Déclaration de conflit d'intérêts
Declaration of competing interest The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.