2020
DOI: 10.1109/comst.2020.3007787
|View full text |Cite
|
Sign up to set email alerts
|

Communication-Efficient Edge AI: Algorithms and Systems

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
164
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
2
1

Relationship

1
8

Authors

Journals

citations
Cited by 329 publications
(164 citation statements)
references
References 126 publications
0
164
0
Order By: Relevance
“…In the last years, we have seen the rise of powerful edge devices that provide the computational power to run AI algorithms at the edge, near to the data sources and in realtime [59]. Since algorithms running on the edge can provide a fast response to condition change and also reduce the size of data to be sent to the cloud by implementing preprocessing methods, and AI algorithms running in the cloud can provide optimization, prediction, and planning functions.…”
Section: Hpc and Processing Infrastructuresmentioning
confidence: 99%
“…In the last years, we have seen the rise of powerful edge devices that provide the computational power to run AI algorithms at the edge, near to the data sources and in realtime [59]. Since algorithms running on the edge can provide a fast response to condition change and also reduce the size of data to be sent to the cloud by implementing preprocessing methods, and AI algorithms running in the cloud can provide optimization, prediction, and planning functions.…”
Section: Hpc and Processing Infrastructuresmentioning
confidence: 99%
“…Currently, model inferences are mostly performed in the cloud; but as the diversity of DNN applications grows, other alternatives to the centralized training and inference strategy are required to lessen the burden on network infrastructures [12]. To that, we opt for Edge Computing [19]: a distributed computing paradigm where software-defined networks are built to decentralize data and provide results expected to be the same as which of cloud computing [20]. Solving these problems, edge DNN aims to process DNN models directly on edge devices.…”
Section: Related Work 21 Edge-based Vs Cloud-based Aimentioning
confidence: 99%
“…Solving these problems, edge DNN aims to process DNN models directly on edge devices. However, to utilize their innate computing power, edge computing faces more resource allocation problems, due to the inherent difference in hardware architecture, the need to sample inputs from built-in peripherals, bandwidth constraints and more [18,20]. To offset these problems, deployment on edge-devices requires DNN models to be minimal enough for fast inference and low-latency updates.…”
Section: Related Work 21 Edge-based Vs Cloud-based Aimentioning
confidence: 99%
“…In contrast, the edge-only inference incurs excessive communication overhead caused by the large volume of input data. Fortunately, the device-edge co-inference paradigm [10,11], which forwards an intermediate feature to the edge server for processing, is a promising candidate to reduce the inference latency by striking a balance between the computation and communication overhead.…”
Section: Introductionmentioning
confidence: 99%