2020
DOI: 10.1109/tc.2019.2954495
|View full text |Cite
|
Sign up to set email alerts
|

Grow and Prune Compact, Fast, and Accurate LSTMs

Abstract: Long short-term memory (LSTM) has been widely used for sequential data modeling. Researchers have increased LSTM depth by stacking LSTM cells to improve performance. This incurs model redundancy, increases run-time delay, and makes the LSTMs more prone to overfitting. To address these problems, we propose a hidden-layer LSTM (H-LSTM) that adds hidden layers to LSTM's original onelevel non-linear control gates. H-LSTM increases accuracy while employing fewer external stacked layers, thus reducing the number of … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
55
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3

Relationship

1
5

Authors

Journals

citations
Cited by 79 publications
(55 citation statements)
references
References 46 publications
0
55
0
Order By: Relevance
“…Network growth is a complementary method to pruning that enables a sparser, yet more accurate, model before pruning starts [10], [27]. A grow-and-prune synthesis paradigm typically reduces the number of parameters in CNNs [10], [28] and LSTMs [29] by another 2×, and increases the classification accuracy [10]. It enables NN based inference even on Internet-of-Things (IoT) sensors [28].…”
Section: Efficient Neural Networkmentioning
confidence: 99%
See 4 more Smart Citations
“…Network growth is a complementary method to pruning that enables a sparser, yet more accurate, model before pruning starts [10], [27]. A grow-and-prune synthesis paradigm typically reduces the number of parameters in CNNs [10], [28] and LSTMs [29] by another 2×, and increases the classification accuracy [10]. It enables NN based inference even on Internet-of-Things (IoT) sensors [28].…”
Section: Efficient Neural Networkmentioning
confidence: 99%
“…Then, it prunes away insignificant connections and neurons based on magnitude information to drastically reduce model redundancy. This leads to improved accuracy and efficiency [10], [29], where the former is highly preferred on the server and the latter is critical at the edge. The training process generates two inference models, i.e., DiabNN-server and DiabNN-edge, for server and edge inference, respectively.…”
Section: Model Trainingmentioning
confidence: 99%
See 3 more Smart Citations