2022
DOI: 10.1162/tacl_a_00513
|View full text |Cite
|
Sign up to set email alerts
|

Robust Dialogue State Tracking with Weak Supervision and Sparse Data

Abstract: Generalizing dialogue state tracking (DST) to new data is especially challenging due to the strong reliance on abundant and fine-grained supervision during training. Sample sparsity, distributional shift, and the occurrence of new concepts and topics frequently lead to severe performance degradation during inference. In this paper we propose a training strategy to build extractive DST models without the need for fine-grained manual span labels. Two novel input-level dropout methods mitigate the negative impact… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(9 citation statements)
references
References 37 publications
0
9
0
Order By: Relevance
“…As shown in Figure 1, experimental results on the MultiWOZ benchmark (Budzianowski et al, 2018) represent a significant milestone. Our approach is the first that, without further fine-tuning, enables modestly sized open-source LLMs (7B or 13B parameters) to achieve comparable or superior performance compared to previous state-of-the-art (SOTA) prompting methods that relied exclusively on advanced proprietary LLMs such as ChatGPT and Codex (Hudeček and Dušek, 2023;Heck et al, 2023;. Furthermore, our approach beats the previous zero-shot SOTA by 5.6% Av.…”
Section: Zero-shot Dst Paradigmsmentioning
confidence: 79%
See 4 more Smart Citations
“…As shown in Figure 1, experimental results on the MultiWOZ benchmark (Budzianowski et al, 2018) represent a significant milestone. Our approach is the first that, without further fine-tuning, enables modestly sized open-source LLMs (7B or 13B parameters) to achieve comparable or superior performance compared to previous state-of-the-art (SOTA) prompting methods that relied exclusively on advanced proprietary LLMs such as ChatGPT and Codex (Hudeček and Dušek, 2023;Heck et al, 2023;. Furthermore, our approach beats the previous zero-shot SOTA by 5.6% Av.…”
Section: Zero-shot Dst Paradigmsmentioning
confidence: 79%
“…( 2) Previous prompting approaches that have only shown efficacy with advanced ChatGPT and Codex. These include IC-DST (Hu et al, 2022) using Codex, (Heck et al, 2023) and InstructTODS using ChatGPT (GPT-3.5/4).…”
Section: Methodsmentioning
confidence: 99%
See 3 more Smart Citations