Energy-Aware Inference Offloading for DNN-Driven Applications in Mobile Edge Clouds

Research output: Journal Publications and Reviews (RGC: 21, 22, 62)21_Publication in refereed journalpeer-review

4 Scopus Citations
View graph of relations

Author(s)

  • Zichuan Xu
  • Liqian Zhao
  • Omer F. Rana
  • Pan Zhou
  • Qiufen Xia
  • Wenzheng Xu
  • Guowei Wu

Detail(s)

Original languageEnglish
Article number9234011
Pages (from-to)799-814
Journal / PublicationIEEE Transactions on Parallel and Distributed Systems
Volume32
Issue number4
Online published20 Oct 2020
Publication statusPublished - Apr 2021
Externally publishedYes

Abstract

With increasing focus on Artificial Intelligence (AI) applications, Deep Neural Networks (DNNs) have been successfully used in a number of application areas. As the number of layers and neurons in DNNs increases rapidly, significant computational resources are needed to execute a learned DNN model. This ever-increasing resource demand of DNNs is currently met by large-scale data centers with state-of-the-art GPUs. However, increasing availability of mobile edge computing and 5G technologies provide new possibilities for DNN-driven AI applications, especially where these application make use of data sets that are distributed in different locations. One fundamental process of a DNN-driven application in mobile edge clouds is the adoption of 'inferencing' - the process of executing a pre-trained DNN based on newly generated image and video data from mobile devices. We investigate offloading DNN inference requests in a 5G-enabled mobile edge cloud (MEC), with the aim to admit as many inference requests as possible. We propose exact and approximate solutions to the problem of inference offloading in MECs. We also consider dynamic task offloading for inference requests, and devise an online algorithm that can be adapted in real time. The proposed algorithms are evaluated through large-scale simulations and using a real world test-bed implementation. The experimental results demonstrate that the empirical performance of the proposed algorithms outperform their theoretical counterparts and other similar heuristics reported in literature.

Research Area(s)

  • approximation and online algorithms, Inference offloading, mobile edge clouds

Citation Format(s)

Energy-Aware Inference Offloading for DNN-Driven Applications in Mobile Edge Clouds. / Xu, Zichuan; Zhao, Liqian; Liang, Weifa; Rana, Omer F.; Zhou, Pan; Xia, Qiufen; Xu, Wenzheng; Wu, Guowei.

In: IEEE Transactions on Parallel and Distributed Systems, Vol. 32, No. 4, 9234011, 04.2021, p. 799-814.

Research output: Journal Publications and Reviews (RGC: 21, 22, 62)21_Publication in refereed journalpeer-review