2020
DOI: 10.48550/arxiv.2005.00891
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Zero-Shot Transfer Learning with Synthesized Data for Multi-Domain Dialogue State Tracking

Abstract: Zero-shot transfer learning for multi-domain dialogue state tracking can allow us to handle new domains without incurring the high cost of data acquisition. This paper proposes new zero-short transfer learning technique for dialogue state tracking where the in-domain training data are all synthesized from an abstract dialogue model and the ontology of the domain. We show that data augmentation through synthesized data can improve the accuracy of zero-shot learning for both the TRADE model and the BERT-based SU… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
15
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 10 publications
(15 citation statements)
references
References 12 publications
0
15
0
Order By: Relevance
“…To address this issue, several methods Zhou and Small, 2019) have proposed for transferring prior knowledge of existing domains to new ones. On the other hand, Campagna et al (2020) proposed an abstract dialogue model that leverages the ontology and in-domain templates to generate a large amount of synthesized data for domain adaptation. Different from their method, in this paper, we utilize a pre-trained seq2seq model and slot descriptions for cross-domain DST without any in-domain data.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…To address this issue, several methods Zhou and Small, 2019) have proposed for transferring prior knowledge of existing domains to new ones. On the other hand, Campagna et al (2020) proposed an abstract dialogue model that leverages the ontology and in-domain templates to generate a large amount of synthesized data for domain adaptation. Different from their method, in this paper, we utilize a pre-trained seq2seq model and slot descriptions for cross-domain DST without any in-domain data.…”
Section: Related Workmentioning
confidence: 99%
“…Note that the reported averaged zero shot joint goal accuracy is not comparable to multi-domains joint goal accuracy. *Result from(Campagna et al, 2020).…”
mentioning
confidence: 99%
“…Training a DST model often requires extensive annotated dialogue data. These data are often collected via a Wizard-of-Oz (Woz) (Kelley, 1984) setting, where two workers converse with each other and annotate the dialogue states of each utterance (Wen et al, 2017;Budzianowski et al, 2018;, or with a Machines Talking To Machines * Work done during internship at Facebook (M2M) framework (Shah et al, 2018), where dialogues are synthesized via the system and user simulators (Campagna et al, 2020;Lin et al, 2021b). However, both of these approaches have inherent challenges when scaling to large datasets.…”
Section: Introductionmentioning
confidence: 99%
“… adapt a copy mechanism for transferring prior knowledge of existing domains to new ones, whileZhou and Small (2019) use the ontology graph to facilitate domain knowledge transfer Campagna et al (2020). leverage the ontology and in-domain templates to generate a large amount of synthesized data for domain adaptation, and apply schema descriptions for tracking out-of-domain slots.…”
mentioning
confidence: 99%
See 1 more Smart Citation