2021
DOI: 10.48550/arxiv.2108.00568
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

FLASH: Fast Neural Architecture Search with Hardware Optimization

Abstract: Neural architecture search (NAS) is a promising technique to design efficient and high-performance deep neural networks (DNNs). As the performance requirements of ML applications grow continuously, the hardware accelerators start playing a central role in DNN design. This trend makes NAS even more complicated and time-consuming for most real applications. This paper proposes FLASH, a very fast NAS methodology that co-optimizes the DNN accuracy and performance on a real hardware platform. As the main theoretica… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 53 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?