Proceedings of the 2019 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays 2019
DOI: 10.1145/3289602.3293904
|View full text |Cite
|
Sign up to set email alerts
|

Req-Yolo

Abstract: Deep neural networks (DNNs), as the basis of object detection, will play a key role in the development of future autonomous systems with full autonomy. The autonomous systems have special requirements of real-time, energy-efficient implementations of DNNs on a power-constrained system. Two research thrusts are dedicated to performance and energy efficiency enhancement of the inference phase of DNNs. The first one is model compression techniques while the second is efficient hardware implementation. Recent work… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
12
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 83 publications
(12 citation statements)
references
References 48 publications
0
12
0
Order By: Relevance
“…Is it possible to further optimize this detector for our case of binary images? One popular technique for optimizing 2D deep networks is to quantize floating point operations to integer operations [46], [47]. The weights of 2D deep networks are generally stored as 32-bit floating point numbers (float32).…”
Section: Generalization Across Unknown Experimental Conditionsmentioning
confidence: 99%
See 1 more Smart Citation
“…Is it possible to further optimize this detector for our case of binary images? One popular technique for optimizing 2D deep networks is to quantize floating point operations to integer operations [46], [47]. The weights of 2D deep networks are generally stored as 32-bit floating point numbers (float32).…”
Section: Generalization Across Unknown Experimental Conditionsmentioning
confidence: 99%
“…Converting these weights to 16-bit floating point numbers (float16) or 8-bit integers (int8) provides memory efficiency, reduces inference latency, and allows leveraging embedded platforms [46]. However, this is achieved at the cost of reduced performance [47].…”
Section: Generalization Across Unknown Experimental Conditionsmentioning
confidence: 99%
“…this detector for our case of binary images? One popular technique for optimizing 2D deep networks is to quantize floating point operations to integer operations [46], [47]. The weights of 2D deep networks are generally stored as 32-bit floating point numbers (float32).…”
Section: Generalization Across Unknown Experimental Conditionsmentioning
confidence: 99%
“…There have been several accelerator architecture designs for low-precision CNNs with uniform quantization arithmetic. Recent literature includes commercial architectures [22], [23] and also academic approaches [24]- [28]. The benefits, in terms of power and throughput, of fitting a design on-chip was Example of a reconfigurable multiplier with the coefficient set {12305, 20746}.…”
Section: Related Workmentioning
confidence: 99%