2019
DOI: 10.1109/tcsi.2019.2895216
|View full text |Cite
|
Sign up to set email alerts
|

Recursive Binary Neural Network Training Model for Efficient Usage of On-Chip Memory

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 43 publications
0
5
0
Order By: Relevance
“…We evaluate the performance of EILE by training a fully-connected network with 2 hidden layers (network size: 784-512-256-10, total 1 MB of parameters) on the full MNIST [13] handwritten digit dataset. Activations are quantized to fixed-point Q (8,8) format while weights and gradients are quantized to Q (2,14) format for batch size of 1, where Q(m, n) denotes the quantization using m bits for the integer part and n bits for the fraction.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…We evaluate the performance of EILE by training a fully-connected network with 2 hidden layers (network size: 784-512-256-10, total 1 MB of parameters) on the full MNIST [13] handwritten digit dataset. Activations are quantized to fixed-point Q (8,8) format while weights and gradients are quantized to Q (2,14) format for batch size of 1, where Q(m, n) denotes the quantization using m bits for the integer part and n bits for the fraction.…”
Section: Resultsmentioning
confidence: 99%
“…Other implementations that support on-chip training on the edge include an FPGA design [7] that uses a special memory management unit to alleviate the impact of irregular memory accesses but the performance during BP is still below that of the FP. A recent study [8] exploits a recursive algorithm for training binary neural networks; however, the processing element (PE) utilization efficiency of FP and BP phases were not reported. Other implementations [9], [10], [11] use custom PE architectures to support on-chip training of different DNN architectures, but their performance either decreases during BP [9] or decreases with smaller batch sizes [10], [11].…”
Section: Introductionmentioning
confidence: 99%
“…FINN [38] proposes a framework for binary neural networks inference on FPGA. ReBNN [10] focuses on reducing memory usage when training BNN. WRPN [30] synthesizes an ASIC for multiple precisions including binary.…”
Section: Related Workmentioning
confidence: 99%
“…Figure. 3 shows the block diagram of the proposed VTC circuit design. It consists of a sampling circuit, an inverter, and a current source.…”
Section: B Proposed Voltage-to-time Converter (Vtc)mentioning
confidence: 99%
“…Recently, research has been focused on AI applications to address complex machine learning problems such as image/speech recognition and language translation [2]. Deep neural networks (DNNs) are widely utilized in such applications since it can achieve high accuracy [3]. However, DNN algorithms are computationally intensive, with large data sets that require high memory bandwidth.…”
Section: Introductionmentioning
confidence: 99%