2020
DOI: 10.1109/tvlsi.2020.2995135
|View full text |Cite
|
Sign up to set email alerts
|

NeuPart: Using Analytical Models to Drive Energy-Efficient Partitioning of CNN Computations on Cloud-Connected Mobile Clients

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 12 publications
(4 citation statements)
references
References 27 publications
0
4
0
Order By: Relevance
“…Typically, publications in this topic discuss the problem of deploying AI models in a real scenario or in a scenario with peculiar constraints that challenge a standard approach. For example, deployment publications showcase the challenges of deploying energy‐efficient AI in FPGA (Tao et al, 2020), in Edge devices (Gondi & Pratap, 2021; Kim & Wu, 2020), and in mobile devices (Jayakodi et al, 2020; Manasi et al, 2020; Wang et al, 2022), and so on.…”
Section: Resultsmentioning
confidence: 99%
“…Typically, publications in this topic discuss the problem of deploying AI models in a real scenario or in a scenario with peculiar constraints that challenge a standard approach. For example, deployment publications showcase the challenges of deploying energy‐efficient AI in FPGA (Tao et al, 2020), in Edge devices (Gondi & Pratap, 2021; Kim & Wu, 2020), and in mobile devices (Jayakodi et al, 2020; Manasi et al, 2020; Wang et al, 2022), and so on.…”
Section: Resultsmentioning
confidence: 99%
“…where D tot (t) is defined through (7), (9), and (10), while E tot is defined through (8) and (11). Under feasibility assumption of (12), and i.i.d.…”
Section: A Proposed Solutionmentioning
confidence: 99%
“…In [7], the authors propose a strategy to partition the DNN under different network conditions to minimize the overall processing delay. Also, [8] minimizes the energy consumption on the client-side by partitioning Convolutional Neural Network (CNN) computations between the client and the cloud. Mao et al partition the DNN always after the first convolutional layer to minimize the cost of mobile devices and use the differentially private mechanism to preserve the privacy [9].…”
Section: Introductionmentioning
confidence: 99%
“…Considered a major drawback, the "high computational complexity" with respect to compute explanations for a single prediction demands "minutes," even for the state-of-the-art model of LIME, GoogleNet (Samek & Müller, 2019). Although efforts are in place attempting to reduce energy consumption (Manasi et al, 2020), these types of models are still characterized by "high energy consumption" (Yang et al, 2017).…”
Section: As Explanation Messenger and Elucidatormentioning
confidence: 99%