2023
DOI: 10.1109/access.2023.3320133
|View full text |Cite
|
Sign up to set email alerts
|

Differentiable Neural Architecture, Mixed Precision and Accelerator Co-Search

Krishna Teja Chitty-Venkata,
Yiming Bian,
Murali Emani
et al.

Abstract: Quantization, effective Neural Network architecture, and efficient accelerator hardware are three important design paradigms to maximize accuracy and efficiency. Mixed Precision Quantization is a process of assigning different precision to different Neural Network layers for optimized inference. Neural Architecture Search (NAS) is a process of automatically designing the neural network for a task and can also be extended to search for the precision of each weight and activation matrix. In this paper, we develo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
references
References 45 publications
0
0
0
Order By: Relevance