2020 International Conference on Information Science and Education (ICISE-IE) 2020
DOI: 10.1109/icise51755.2020.00107
|View full text |Cite
|
Sign up to set email alerts
|

A Graph Signal Processing Based Strategy for Deep Neural Network Inference

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(9 citation statements)
references
References 7 publications
0
9
0
Order By: Relevance
“…Several studies exist for operator execution order scheduling, such as [2,38,39]. Among these efforts [38,39] focus on minimizing peak memory consumption by reordering operators for resource-constrained devices (e.g., MCUs), and effort [2] proposes an optimized scheduling framework for complex models (irregularly wired neural networks). These approaches rely on static shapes only.…”
Section: Related Workmentioning
confidence: 99%
“…Several studies exist for operator execution order scheduling, such as [2,38,39]. Among these efforts [38,39] focus on minimizing peak memory consumption by reordering operators for resource-constrained devices (e.g., MCUs), and effort [2] proposes an optimized scheduling framework for complex models (irregularly wired neural networks). These approaches rely on static shapes only.…”
Section: Related Workmentioning
confidence: 99%
“…Several techniques have been developed to address TinyML's low-resource challenges, including pruning [39], [40], [41], [42], [43], [44], [45], Quantization [46], [47], [48], [49], [50], [39], [51], [52], [53], [54], [55] and neural architecture search (NAS) [53], [56], [57], [58], [59], [60], [61], [62], [63], [64]. These methods reduce model parameters while maintaining model accuracy, allowing the models to be applied to MCUs.…”
Section: Aifesmentioning
confidence: 99%
“…By using a similar approach, DNNs were split between edge (Jetson TX2 and NVIDIA Drive PX2 devices) and cloud domains [22]. Finally, memory efficient patch-based inference for microcontrollers (with only hundreds KBs of RAM) has been proposed [23], which reduces peak memory consumption of existing models by 4-8x. Google has published several convolutional neural networks (CNNs) in Tensorflow Hub for human joint keypoint identification [13].…”
Section: Literature Reviewmentioning
confidence: 99%