Interspeech 2018 2018
DOI: 10.21437/interspeech.2018-1057
|View full text |Cite
|
Sign up to set email alerts
|

Structured Word Embedding for Low Memory Neural Network Language Model

Abstract: Neural network language model (NN LM), such as long short term memory (LSTM) LM, has been increasingly popular due to its promising performance. However, the model size of an uncompressed NN LM is still too large to be used in embedded or portable devices. The dominant part of memory consumption of NN LM is the word embedding matrix. Directly compressing the word embedding matrix usually leads to performance degradation. In this paper, a product quantization based structured embedding approach is proposed to s… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
18
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 15 publications
(18 citation statements)
references
References 16 publications
0
18
0
Order By: Relevance
“…However, running these models on edge-devices, faces memory and latency issues due to limitations of the hardware. Thus, there has been considerable interest towards research in reducing the memory footprint and faster inference speed for these models (Sainath et al, 2013;Acharya et al, 2019;Shi and Yu, 2018;Jegou et al, 2010;Chen et al, 2018;Winata et al, 2019).…”
Section: Introductionmentioning
confidence: 99%
See 4 more Smart Citations
“…However, running these models on edge-devices, faces memory and latency issues due to limitations of the hardware. Thus, there has been considerable interest towards research in reducing the memory footprint and faster inference speed for these models (Sainath et al, 2013;Acharya et al, 2019;Shi and Yu, 2018;Jegou et al, 2010;Chen et al, 2018;Winata et al, 2019).…”
Section: Introductionmentioning
confidence: 99%
“…Recently, there has been considerable work on compressing word-embedding matrices (Sainath et al, 2013;Acharya et al, 2019;Shi and Yu, 2018;Jegou et al, 2010;Chen et al, 2018;Winata et al, 2019). These techniques have proven to perform atpar with the uncompressed models, but still suffer from a number of issues.…”
Section: Introductionmentioning
confidence: 99%
See 3 more Smart Citations