I. Introduction
The rapid increase of Internet of Things (IoT) devices has raised the demand for IoT-enabled artificial intelligence (AI) platforms. Located in the cloud, these AI platforms learn a new capability from the existing data (training) and apply this capability to new data via an application or service (inference). However, these AI platforms require high computing power to support deep learning (DL) applications, diverse computing resources, massive networking, and high response latency [1]. These challenges can be met by pooling the computing resources at the edge of the network, a concept known as edge computing (EC) [2]. Here, an edge refers to a location, far from the cloud, where an edge device runs edge applications. EC defines the operation of running workloads on these devices [2].