2021 IEEE 14th International Conference on Cloud Computing (CLOUD) 2021
DOI: 10.1109/cloud53861.2021.00016
|View full text |Cite
|
Sign up to set email alerts
|

AI Multi-Tenancy on Edge: Concurrent Deep Learning Model Executions and Dynamic Model Placements on Edge Devices

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
11
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3

Relationship

3
5

Authors

Journals

citations
Cited by 16 publications
(11 citation statements)
references
References 20 publications
0
11
0
Order By: Relevance
“…Only a few solutions have been designed for GPU management in the context of edge computing. For example, Subedi et al [39] mainly focuses on enabling GPU accelerated edge computation without considering latency-critical aspects such as placing applications close to the edge clients.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Only a few solutions have been designed for GPU management in the context of edge computing. For example, Subedi et al [39] mainly focuses on enabling GPU accelerated edge computation without considering latency-critical aspects such as placing applications close to the edge clients.…”
Section: Related Workmentioning
confidence: 99%
“…In addition, the tasks at the edge, like AI-based computations, are becoming heavier and heavier. The use of GPUs [9] can be fundamental to accelerate these computations, but they are seldom taken into account explicitly [6,20,39].…”
mentioning
confidence: 99%
“…• We investigate the beneits and limitations of deploying DL models on edge devices with EdgeTPU clusters. This work is based on our preliminary version [65] and we take one step in a broader and more thorough evaluation of techniques for maximizing the DL inference throughput on edge devices and AI accelerators. In particular, to the best of our knowledge, this work is the irst study on evaluating DMP on heterogeneous edge resources/EdgeTPUs and characterizing the performance of EdgeTPU cluster for DL inference tasks.…”
Section: Introductionmentioning
confidence: 99%
“…The edge computing paradigm is emerging because the edge cluster can be deployed near data sources, e.g., IoT sensors, and process the data without relying on traditional data-center computing [4]- [6]. In particular, the development of small, energy-efficient, and capable CPU and AI accelerators for edge devices facilitates the adoption of edge computing to servicing various real-world applications, e.g., autonomous driving, drone-based surveillance systems, and environmental sensing [7]- [10].…”
Section: Introductionmentioning
confidence: 99%
“…IoT sensor and device layer ("bottom layer" in Fig. 1) has various IoT sensors and user devices, which perform diverse sensing operations, and user devices (including edge devices) are used to host lightweight applications to perform real-time on-board processing, including sensing data filtering and noise removal [53], lightweight AI inference tasks [10], [54], and stream processing [55].…”
Section: Introductionmentioning
confidence: 99%