Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics 2023
DOI: 10.18653/v1/2023.eacl-main.107
|View full text |Cite
|
Sign up to set email alerts
|

Selective In-Context Data Augmentation for Intent Detection using Pointwise V-Information

Yen-Ting Lin,
Alexandros Papangelis,
Seokhwan Kim
et al.

Abstract: This work focuses on in-context data augmentation for intent detection. Having found that augmentation via in-context prompting of large pretrained language models (PLMs) alone does not improve performance, we introduce a novel approach based on PLMs and pointwise Vinformation (PVI), a metric that can measure the usefulness of a datapoint for training a model. Our method first fine-tunes a PLM on a small seed of training data and then synthesizes new datapoints -utterances that correspond to given intents. It … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 34 publications
(49 reference statements)
0
4
0
Order By: Relevance
“…Baseline For intent classification, we compare with various few-shot fine-tuned baselines: RoBERTa (Liu et al, 2019), ICDA (Lin et al, 2023), DNNC (Zhang et al, 2020a), and CPFT (Zhang et al, 2021). While for zeroshot baseline, we employ MNLI (Williams et al, 2018) fine-tuned BART large models (Lewis et al, 2020) by framing intent classification as an NLI task.…”
Section: Experiments Settingsmentioning
confidence: 99%
“…Baseline For intent classification, we compare with various few-shot fine-tuned baselines: RoBERTa (Liu et al, 2019), ICDA (Lin et al, 2023), DNNC (Zhang et al, 2020a), and CPFT (Zhang et al, 2021). While for zeroshot baseline, we employ MNLI (Williams et al, 2018) fine-tuned BART large models (Lewis et al, 2020) by framing intent classification as an NLI task.…”
Section: Experiments Settingsmentioning
confidence: 99%
“…It is used as a quality estimate of Universal Dependencies treebanks in Kulmizev and Nivre (2023). PVI is used to select synthetic data as an augmentation to an intent detection classifier, which achieves state-of-the-art performance (Lin et al, 2023). Chen et al (2022a) and Prasad et al (2023) incorporate PVI into an informativeness metric to evaluate rationales, and find that it captures the expected flow of information in high-quality reasoning chains.…”
Section: Hardnessmentioning
confidence: 99%
“…PVI measures the amount of usable information in an input for a given model, which reflects the ease with which a model can predict a certain label given an input. Though it is a recently proposed method, the effectiveness of PVI has been demonstrated in various NLP tasks (Chen et al, 2022a;Kulmizev and Nivre, 2023;Lin et al, 2023;Prasad et al, 2023).…”
Section: Introductionmentioning
confidence: 99%
“…Task As a fundamental element in task-oriented dialog systems, intent detection is normally conducted in the NLU component for identifying a user's intent given an utterance (Ham et al, 2020). Recently, accurately identifying intents in the fewshot setting has attracted much attention due to data scarcity issues resulted from the cost of data collection as well as privacy and ethical concerns (Lin et al, 2023b). Following the few-shot intent detection benchmark , we focus on the challenging 5-shot and 10-shot settings.…”
Section: Few-shot Intent Classificationmentioning
confidence: 99%