TY - GEN
T1 - Energy-Efficient AI over a Virtualized Cloud Fog Network
AU - Yosuf, Barzan A.
AU - Mohamed, Sanaa H.
AU - Alenazi, Mohammed M.
AU - El-Gorashi, Taisir E.H.
AU - Elmirghani, Jaafar M.H.
N1 - Funding Information:
The authors would like to acknowledge funding from the Engineering and Physical Sciences Research Council (EPSRC), in part by INTERNET under Grant EP/H040536/1, in part by SwiTching And tRansmission (STAR) under Grant EP/K016873/1, and in part by Terabit Bidirectional Multi-user Optical Wireless System (TOWS) for 6G LiFi under Grant EP/S016570/1. All data is provided in the results section of this paper.
Publisher Copyright:
© 2021 ACM.
PY - 2021/6/22
Y1 - 2021/6/22
N2 - Deep Neural Networks (DNNs) have served as a catalyst in introducing a plethora of next-generation services in the era of Internet of Things (IoT), thanks to the availability of massive amounts of data collected by the objects on the edge. Currently, DNN models are used to deliver many Artificial Intelligence (AI) services that include image and natural language processing, speech recognition, and robotics. Accordingly, such services utilize various DNN models that make it computationally intensive for deployment on the edge devices alone. Thus, most AI models are offloaded to distant cloud data centers (CDCs), which tend to consolidate large amounts of computing and storage resources into one or more CDCs. Deploying services in the CDC will inevitably lead to excessive latencies and overall increase in power consumption. Instead, fog computing allows for cloud services to be extended to the edge of the network, which allows for data processing to be performed closer to the end-user device. However, different from cloud data centers, fog nodes have limited computational power and are highly distributed in the network. In this paper, using Mixed Integer Linear Programming (MILP), we formulate the placement of DNN inference models, which is abstracted as a network embedding problem in a Cloud Fog Network (CFN) architecture, where power savings are introduced through trade-offs between processing and networking. We study the performance of the CFN architecture by comparing the energy savings when compared to the baseline approach which is the CDC.
AB - Deep Neural Networks (DNNs) have served as a catalyst in introducing a plethora of next-generation services in the era of Internet of Things (IoT), thanks to the availability of massive amounts of data collected by the objects on the edge. Currently, DNN models are used to deliver many Artificial Intelligence (AI) services that include image and natural language processing, speech recognition, and robotics. Accordingly, such services utilize various DNN models that make it computationally intensive for deployment on the edge devices alone. Thus, most AI models are offloaded to distant cloud data centers (CDCs), which tend to consolidate large amounts of computing and storage resources into one or more CDCs. Deploying services in the CDC will inevitably lead to excessive latencies and overall increase in power consumption. Instead, fog computing allows for cloud services to be extended to the edge of the network, which allows for data processing to be performed closer to the end-user device. However, different from cloud data centers, fog nodes have limited computational power and are highly distributed in the network. In this paper, using Mixed Integer Linear Programming (MILP), we formulate the placement of DNN inference models, which is abstracted as a network embedding problem in a Cloud Fog Network (CFN) architecture, where power savings are introduced through trade-offs between processing and networking. We study the performance of the CFN architecture by comparing the energy savings when compared to the baseline approach which is the CDC.
KW - cloud-fog networks
KW - Deep Neural Network (DNN) placement
KW - energy efficiency
KW - Internet-of-Things) IoT
KW - Mixed Integer Linear Programming (MILP)
KW - optimization
KW - resource allocation
UR - http://www.scopus.com/inward/record.url?scp=85109305673&partnerID=8YFLogxK
UR - https://energy.acm.org/conferences/eenergy/2021/
U2 - 10.1145/3447555.3465378
DO - 10.1145/3447555.3465378
M3 - Conference contribution
AN - SCOPUS:85109305673
T3 - e-Energy 2021 - Proceedings of the 2021 12th ACM International Conference on Future Energy Systems
SP - 328
EP - 334
BT - Proceedings of the 202112th ACM International Conference on Future Energy Systems
A2 - De Meer, Hermann
A2 - Meo, Michela
A2 - Ardakanian, Omid
A2 - Nieße, Astrid
PB - Association for Computing Machinery (ACM)
CY - United States
T2 - 12th ACM International Conference on Future Energy Systems, e-Energy 2021
Y2 - 28 June 2021 through 2 July 2021
ER -