Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.620
|View full text |Cite
|
Sign up to set email alerts
|

Knowledge-Aware Graph-Enhanced GPT-2 for Dialogue State Tracking

Abstract: Dialogue State Tracking is central to multidomain task-oriented dialogue systems, responsible for extracting information from user utterances. We present a novel hybrid architecture that augments GPT-2 with representations derived from Graph Attention Networks in such a way to allow causal, sequential prediction of slot values. The model architecture captures inter-slot relationships and dependencies across domains that otherwise can be lost in sequential prediction. We report improvements in state tracking pe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 22 publications
(13 citation statements)
references
References 4 publications
0
13
0
Order By: Relevance
“…Another benefit of using pretrained GPT2 is faster training time as we observed the VDTN+GPT2 converged much earlier than training it from scratch. From these observations, we are excited to see more future extension of SOTA unimodal DST models (Lin et al, 2021;Dai et al, 2021) and large pretrained LMs (Brown et al, 2020;Raffel et al, 2020), especially ones with multimodal learning such as (Lu et al, 2019;, to MM-DST task.…”
Section: Resultsmentioning
confidence: 99%
“…Another benefit of using pretrained GPT2 is faster training time as we observed the VDTN+GPT2 converged much earlier than training it from scratch. From these observations, we are excited to see more future extension of SOTA unimodal DST models (Lin et al, 2021;Dai et al, 2021) and large pretrained LMs (Brown et al, 2020;Raffel et al, 2020), especially ones with multimodal learning such as (Lu et al, 2019;, to MM-DST task.…”
Section: Resultsmentioning
confidence: 99%
“…In a multi-domain dialog, slots from different domains can be correlated. The slot relationships can be modeled implicitly through self-attention or graph neural networks (Kim et al, 2020;Ye et al, 2021;Lin et al, 2021a), or explicitly through copy mechanism or a schema graph that encodes prior knowledge (Heck et al, 2020;Jiao et al, 2022;Feng et al, 2022). However, all of these methods are driven by multi-domain data.…”
Section: Multi-domain Dialog State Trackingmentioning
confidence: 99%
“…In addition, we observe that the complete states of the dialogue session are updated at turn-8, while turn-9 and turn-10 simply show humans' politeness and respect without introducing any new domain-slots. Therefore, while the "last turn" has been studied before (Lin et al, 2021a), it is often not the case that only the last turn of a dialogue session generates summary states. Using redundant turns such as turn-9 and turn-10 for training not only requires additional labelling but also possibly distracts the DST model since it introduces irrelevant context information, thus hindering the overall performance .…”
Section: Will You Need Any Thing Else Now?mentioning
confidence: 99%