2017 IEEE International Conference on Cluster Computing (CLUSTER) 2017
DOI: 10.1109/cluster.2017.45
|View full text |Cite
|
Sign up to set email alerts
|

A Power-Efficient Accelerator Based on FPGAs for LSTM Network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 24 publications
(10 citation statements)
references
References 5 publications
0
10
0
Order By: Relevance
“…In future studies, a wider range of activities should be included to provide more information about health-related daily PAs. Though we achieved a high classification performance using the RF classifier, applying other advanced machine learning models such as recurrent neural networks including long short-term memory (LSTM) networks [45,46] and comparing their performance may be considered as a future study. Finally, we trained the models using data collected by young healthy adults only.…”
Section: Contributions and Limitationsmentioning
confidence: 99%
“…In future studies, a wider range of activities should be included to provide more information about health-related daily PAs. Though we achieved a high classification performance using the RF classifier, applying other advanced machine learning models such as recurrent neural networks including long short-term memory (LSTM) networks [45,46] and comparing their performance may be considered as a future study. Finally, we trained the models using data collected by young healthy adults only.…”
Section: Contributions and Limitationsmentioning
confidence: 99%
“…Zhang et al (2017) deployed an efficient accelerator which targeted for the LSTM networks execution. They minimized the power consumption, time and energy by pipelining the execution behaviour for matrix multiplication operations, element wise computation etc.…”
Section: Related Workmentioning
confidence: 99%
“…As shown in Table 1, the trained TM exhibits significant performance improvement over the SM trained in the standard cross-entropy loss. Nevertheless, it is worth noting that the training time of the SM is only about a quarter of the TM's training time, and the testing time of the SM is only about one-sixth of the TM's testing time as the LSTM-RNN operation is very time-consuming [30]. What's more, as illustrated in Table 1, the teaching loss makes the SM obtain the performance improvement of 7.21%, which shows the proposed CMDL is effective for the performance boost of a lightweight model.…”
Section: ) Performance Analysis Of Cmdlmentioning
confidence: 99%
“…For example, in [28] the network contains 313,603 parameters with one LSTM-RNN layer, four convolution layers, and two fully connected layers. In addition, the time complexity of the models comprised of CNN and LSTM-RNN layers is high in training or prediction as the LSTM-RNN operation is timeconsuming [30]. Thus, these networks take a long time to automatically predict the types of signals.…”
Section: Introductionmentioning
confidence: 99%