2020 IEEE 40th International Conference on Distributed Computing Systems (ICDCS) 2020
DOI: 10.1109/icdcs47774.2020.00114
|View full text |Cite
|
Sign up to set email alerts
|

PerDNN: Offloading Deep Neural Network Computations to Pervasive Edge Servers

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(10 citation statements)
references
References 33 publications
0
5
0
Order By: Relevance
“…Reducto [45] uses an on-camera filtering technique to filter out frames that do not contain relevant information for the query. Partition-based methods [38,40,78] explore partitioning the DNNs over the edge and cloud to fully utilize the computation resources on both sides. They automatically divide a DNN model into two partitions and deploy the few initial layers to improve inference efficiency.…”
Section: Related Workmentioning
confidence: 99%
“…Reducto [45] uses an on-camera filtering technique to filter out frames that do not contain relevant information for the query. Partition-based methods [38,40,78] explore partitioning the DNNs over the edge and cloud to fully utilize the computation resources on both sides. They automatically divide a DNN model into two partitions and deploy the few initial layers to improve inference efficiency.…”
Section: Related Workmentioning
confidence: 99%
“…It partitions DNN models based on a penalty factor to reduce the uploading overhead; and it is the most popular approach evaluated in practice. e third is PerDNN [23]: it is a recent work on DNN offloading. It uses the GPU statistics of servers to partition DNN models to minimize the execution latency between the client and the edge server.…”
Section: Workloadsmentioning
confidence: 99%
“…(1) In order to completely push the computing tasks of DNN applications to the edge of the network, after studying the challenges that need to be overcome to execute DNN applications in collaboration with multiple edge nodes, we found some problems that may delay the execution of DNN applications, such as concurrency conflict exceptions, network jitters, and deadlocks. [10][11][12][17][18][19][20][21][22][23][24] offload all (i.e., cloud-only) or part (i.e., edgecloud collaboration) of the DNN computation to the cloud server. Neurosurgeon [10] dynamically divides the DNN model into the front part and the rear part.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…In the literature, some recent works exploit MEC for computational offloading of DNN tasks under a single WD setup [11]- [21] or a multi-WD setup [21], [22]. The DNN task offloading and resource allocation schemes are designed to optimize the WD's energy consumption [12]- [14], the DNN inferencing accuracy [15], [16], and the DNN inferencing time [17]- [20]. The profiling knowledge of layer-wise DNN inferencing delay/energy consumption, which heavily depends on the MEC system parameters, is determined in either an offline manner [12]- [18] or by an online learning approach [19], [20].…”
Section: Introductionmentioning
confidence: 99%