2020
DOI: 10.1609/aaai.v34i05.6349
|View full text |Cite
|
Sign up to set email alerts
|

MOSS: End-to-End Dialog System Framework with Modular Supervision

Abstract: A major bottleneck in training end-to-end task-oriented dialog system is the lack of data. To utilize limited training data more efficiently, we propose Modular Supervision Network (MOSS), an encoder-decoder training framework that could incorporate supervision from various intermediate dialog system modules including natural language understanding, dialog state tracking, dialog policy learning and natural language generation. With only 60% of the training data, MOSS-all (i.e., MOSS with supervision from all f… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
28
0

Year Published

2020
2020
2025
2025

Publication Types

Select...
3
3
2

Relationship

1
7

Authors

Journals

citations
Cited by 41 publications
(28 citation statements)
references
References 13 publications
0
28
0
Order By: Relevance
“…One major difference between our Graph Reasoning Module and standard GNN is that, we want the message passing in layer L conditioned on the L th instruction vector. Inspired by language model type condition (Liang et al, 2020b), we adopt a general design that is compatible with any graph neural network design: Before running the L th GNN layer, we concatenate the L th instruction vector to every node and edge feature from the previous layer. Specifically,…”
Section: Graph Reasoning Modulementioning
confidence: 99%
“…One major difference between our Graph Reasoning Module and standard GNN is that, we want the message passing in layer L conditioned on the L th instruction vector. Inspired by language model type condition (Liang et al, 2020b), we adopt a general design that is compatible with any graph neural network design: Before running the L th GNN layer, we concatenate the L th instruction vector to every node and edge feature from the previous layer. Specifically,…”
Section: Graph Reasoning Modulementioning
confidence: 99%
“…We borrow some elements from the Sequicity ) model, such as representing the belief state as a natural language sequence (a text span), and using copy-augmented Seq2Seq learning (Gu et al, 2016). But compared to Sequicity and all its follow-up works Shu et al, 2019;Liang et al, 2020), a feature in our LABES-S2S model is that the transition between belief states across turns and the dependency between system responses and belief states are well statistically modeled. This new design results in a completely different graphical model structure, which enables rigorous probabilistic variational learning.…”
Section: Related Workmentioning
confidence: 99%
“…E2E Models: E2E models can be divided into three sub-categories. The TSCP , SEDST , FSDM (Shu et al, 2019), MOSS (Liang et al, 2020) and DAMD are based on the copy-augmented Seq2Seq learning framework proposed by . LIDM (Wen et al, 2017a), SFN (Mehri et al, 2019) and UniConv (Le et al, 2020a) are modular designed, connected through neural states and trained end-to-end.…”
Section: Baselinesmentioning
confidence: 99%
See 1 more Smart Citation
“…1(b) shows an O2O model with a conditional chain mapping. This method for multiple sequence modeling has been applied to dialog modeling (Liang et al, 2020), speaker diarization (Fujita et al, 2020a), and multi-speaker ASR (Shi et al, 2020). Unlike the O2M model, this model can predict a variable number of output sequences while explicitly considering dependencies between the multiple sequences based on the probabilistic chain rule.…”
Section: Introductionmentioning
confidence: 99%