2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA) 2015
DOI: 10.1109/icmla.2015.121
|View full text |Cite
|
Sign up to set email alerts
|

Model Shrinking for Embedded Keyword Spotting

Abstract: In this paper we present two approaches to improve computational efficiency of a keyword spotting system running on a resource constrained device. This embedded keyword spotting system detects a pre-specified keyword in real time at low cost of CPU and memory. Our system is a two stage cascade. The first stage extracts keyword hypotheses from input audio streams. After the first stage is triggered, hand-crafted features are extracted from the keyword hypothesis and fed to a support vector machine (SVM) classif… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
8
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
6
2
1

Relationship

3
6

Authors

Journals

citations
Cited by 17 publications
(8 citation statements)
references
References 20 publications
0
8
0
Order By: Relevance
“…one-hot vector) without considering them as sequences of characters. Other recent works aim to spot specific keywords used to activate voice assistant systems [29,30,31]. The application of BiLSTMs on KWS was first proposed in [32].…”
Section: Related Workmentioning
confidence: 99%
“…one-hot vector) without considering them as sequences of characters. Other recent works aim to spot specific keywords used to activate voice assistant systems [29,30,31]. The application of BiLSTMs on KWS was first proposed in [32].…”
Section: Related Workmentioning
confidence: 99%
“…Recently, with increasing popularity of voice assistant systems, small-footprint keyword spotting systems have been attracting more attention [4,5,6]. For example, Alexa on Amazon Tap requires a keyword spotting system running continuously on the device under tight CPU, memory, latency and power usage constraints.…”
Section: Introductionmentioning
confidence: 99%
“…Gaussian mixture model (GMM) is a traditional one for modeling observation probabilities of HMM. With deep neural network (DNN) gradually takes the place of GMM in speech recognition, DNN and other neural network based models are also attempted in this structure [7,8]. In recent years, the systems that only use neural network without HMM involved are proposed [9,10].…”
Section: Introductionmentioning
confidence: 99%