Early Exiting (EE) is an emerging computing paradigm where Deep Neural Networks (DNNs) are equipped with earlier classifiers, enabling trading-off accuracy with in- ference latency. EE can be effectively combined with edge computing, a paradigm that allows mobile nodes to offload complex tasks, such as the execution of DNNs, to servers at the edge of the network, thus reducing computing times and energy consumption at the mobile devices. The integration of such technologies is particularly attractive for the support of applications for connected and automated driving. In this paper, we consider a system that jointly leverages the benefits of EE and edge computing, and we model their complex interactions by means of a Markov Decision Process (MDP). We then formulate an optimization problem to select the inference strategy that maximizes the average task accuracy. Importantly, such an optimization problem has low complexity, as the optimal policy can be derived by mapping the MDP into a linear program. Our numerical results focus on a use case centered on automated vehicles connected with an edge server under varying channel and network conditions, and show that our solution achieves up to 11% higher accuracy compared to the optimal policy with no EE.
Edge Computing with Early Exiting for Adaptive Inference in Mobile Autonomous Systems / Angelucci, Simone; Valentini, Roberto; Levorato, Marco; Santucci, Fortunato; Chiasserini, Carla Fabiana. - ELETTRONICO. - (2024), pp. 2080-2085. (Intervento presentato al convegno ICC 2024 - IEEE International Conference on Communications tenutosi a Denver (USA) nel 09-13 June 2024) [10.1109/ICC51166.2024.10622411].
Edge Computing with Early Exiting for Adaptive Inference in Mobile Autonomous Systems
Marco Levorato;Carla Fabiana Chiasserini
2024
Abstract
Early Exiting (EE) is an emerging computing paradigm where Deep Neural Networks (DNNs) are equipped with earlier classifiers, enabling trading-off accuracy with in- ference latency. EE can be effectively combined with edge computing, a paradigm that allows mobile nodes to offload complex tasks, such as the execution of DNNs, to servers at the edge of the network, thus reducing computing times and energy consumption at the mobile devices. The integration of such technologies is particularly attractive for the support of applications for connected and automated driving. In this paper, we consider a system that jointly leverages the benefits of EE and edge computing, and we model their complex interactions by means of a Markov Decision Process (MDP). We then formulate an optimization problem to select the inference strategy that maximizes the average task accuracy. Importantly, such an optimization problem has low complexity, as the optimal policy can be derived by mapping the MDP into a linear program. Our numerical results focus on a use case centered on automated vehicles connected with an edge server under varying channel and network conditions, and show that our solution achieves up to 11% higher accuracy compared to the optimal policy with no EE.File | Dimensione | Formato | |
---|---|---|---|
a495-angelucci final.pdf
accesso aperto
Tipologia:
2. Post-print / Author's Accepted Manuscript
Licenza:
PUBBLICO - Tutti i diritti riservati
Dimensione
424.43 kB
Formato
Adobe PDF
|
424.43 kB | Adobe PDF | Visualizza/Apri |
Edge_Computing_with_Early_Exiting_for_Adaptive_Inference_in_Mobile_Autonomous_Systems.pdf
non disponibili
Tipologia:
2a Post-print versione editoriale / Version of Record
Licenza:
Non Pubblico - Accesso privato/ristretto
Dimensione
512.04 kB
Formato
Adobe PDF
|
512.04 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.
https://hdl.handle.net/11583/2987073