Proceedings of the 28th International Conference on Computational Linguistics 2020
DOI: 10.18653/v1/2020.coling-main.438
|View full text |Cite
|
Sign up to set email alerts
|

Few-shot Pseudo-Labeling for Intent Detection

Abstract: In this paper, we introduce a state-of-the-art pseudo-labeling technique for few-shot intent detection. We devise a folding/unfolding hierarchical clustering algorithm which assigns weighted pseudo-labels to unlabeled user utterances. We show that our two-step method yields significant improvement over existing solutions. This performance is achieved on multiple intent detection datasets, even in more challenging situations where the number of classes is large or when the dataset is highly imbalanced. Moreover… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
4

Relationship

1
9

Authors

Journals

citations
Cited by 13 publications
(6 citation statements)
references
References 22 publications
0
6
0
Order By: Relevance
“…As the important part of a dialog system, dialogue language understanding attract a lot of attention in few-shot scenario. Dopierre et al (2020); Vlasov et al (2018); Xia et al (2018) explored fewshot intent detection technique. Luo et al (2018) and Hou et al (2020a) investigated few-shot slot tagging by using prototypical network.…”
Section: Related Workmentioning
confidence: 99%
“…As the important part of a dialog system, dialogue language understanding attract a lot of attention in few-shot scenario. Dopierre et al (2020); Vlasov et al (2018); Xia et al (2018) explored fewshot intent detection technique. Luo et al (2018) and Hou et al (2020a) investigated few-shot slot tagging by using prototypical network.…”
Section: Related Workmentioning
confidence: 99%
“…Pseudo-label selection strategies. We consider several pseudolabel selection strategies, including uncertainty [37], confidence, and clustering [9]. We fix 𝑢 𝑟 to 0.1 on all datasets.…”
Section: Model Variants (Rq4)mentioning
confidence: 99%
“…They experimentally compare their offered approach to 6 more methods (Matching Network, Prototypical Network, Relation Network, Hybrid Attention-based Prototypical Network, Hierarchical Prototypical Network, Multi-level Matching, and Aggregation Network) and prove their method achieves the best performance on two datasets. A very similar problem [20] is tackled with the novel two-fold pseudolabeling technique. The pseudolabeling process takes embedded user utterances and passes them to a hierarchical clustering method (in a bottom-up tree-manner), then the process goes top-down a tree and expands nodes having multiple labeled sentences with different labels.…”
Section: Related Workmentioning
confidence: 99%