Interspeech 2022 2022
DOI: 10.21437/interspeech.2022-921
|View full text |Cite
|
Sign up to set email alerts
|

Dummy Prototypical Networks for Few-Shot Open-Set Keyword Spotting

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(3 citation statements)
references
References 0 publications
0
3
0
Order By: Relevance
“…In the context of Few-Shot learning for KWS, many works relied on the Prototypical Network (ProtoNet) approach 9 . Similarly to our proposed approach, Pro-toNet uses episodic training and prototype-based classification but cross-entropy is adopted as the optimization criteria.…”
Section: Comparison With Other Workmentioning
confidence: 99%
“…In the context of Few-Shot learning for KWS, many works relied on the Prototypical Network (ProtoNet) approach 9 . Similarly to our proposed approach, Pro-toNet uses episodic training and prototype-based classification but cross-entropy is adopted as the optimization criteria.…”
Section: Comparison With Other Workmentioning
confidence: 99%
“…For example, Huh et al [6] explore several metric learning objectives such as triplet loss [18] and prototypical loss [19] for training KWS. In addition, Kim et al [20] suggest a multiple dummy prototype generator to handle open-set queries efficiently. Considering that KWS is closer to the detection task rather than the classification task in the real-world scenario, metric learning methods are advantageous over classification approaches, effectively tackling unknown category samples via distance metrics.…”
Section: Metric Learning Based Kwsmentioning
confidence: 99%
“…For data augmentation, we used noise corpus from [24] and RIR coefficients from [25]. The maximum value for volume scale is sampled from (0.2, 0.9) and SNR is sampled from (10,20).…”
Section: Implementation Detailsmentioning
confidence: 99%