Interspeech 2021 2021
DOI: 10.21437/interspeech.2021-1966
|View full text |Cite
|
Sign up to set email alerts
|

Few-Shot Keyword Spotting in Any Language

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 16 publications
(5 citation statements)
references
References 11 publications
0
5
0
Order By: Relevance
“…The feature extractor is trained on a source dataset disjoined from the target scenario. We employ the recent Multilingual Spoken Words Corpus (MSWC) dataset 6 , which includes up to 39k unique utterances in the English train partition. The average number of occurrences per class is 180; class distribution is highly unbalanced.…”
Section: Dnn Feature Extractor Designmentioning
confidence: 99%
See 1 more Smart Citation
“…The feature extractor is trained on a source dataset disjoined from the target scenario. We employ the recent Multilingual Spoken Words Corpus (MSWC) dataset 6 , which includes up to 39k unique utterances in the English train partition. The average number of occurrences per class is 180; class distribution is highly unbalanced.…”
Section: Dnn Feature Extractor Designmentioning
confidence: 99%
“…Recently, Mazumder et al 6 used transfer learning to train a DNN-based classifier with few shots. This approach places a linear classifier on top of a backbone model trained on the multilingual MSWC dataset.…”
Section: Comparison With Other Workmentioning
confidence: 99%
“…their encoder and show better generalization ability under fewshot learning. Mazumder et al [10] train a multi-class multilignual keyword classification model with EfficientNet's structure [11] as the encoder on Common Voice [12] to solve multilingual few-shot KWS problem. Nonetheless, preparing largescale KWS datasets usually requires audios, transcription, and a forced aligner, which increases the cost.…”
Section: Train a Multi-class Keyword Classification Model On Librispe...mentioning
confidence: 99%
“…Such classification layers can be independent of each other and use the same embedding model, since in the adaptation phase only the last layer is modified. A similar solution based on a few-shot transfer learning is described in [16]. It should be noted that usually KWS solutions are deployed on devices with limited resources, hence performing any type of model adaptation might be troublesome.…”
Section: Introductionmentioning
confidence: 99%