2019
DOI: 10.1162/tacl_a_00274
|View full text |Cite
|
Sign up to set email alerts
|

Learning End-to-End Goal-Oriented Dialog with Maximal User Task Success and Minimal Human Agent Use

Abstract: Neural end-to-end goal-oriented dialog systems showed promise to reduce the workload of human agents for customer service, as well as reduce wait time for users. However, their inability to handle new user behavior at deployment has limited their usage in real world. In this work, we propose an end-to-end trainable method for neural goal-oriented dialog systems which handles new user behaviors at deployment by transferring the dialog to a human agent intelligently. The proposed method has three goals: 1) maxim… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 10 publications
(12 citation statements)
references
References 14 publications
0
12
0
Order By: Relevance
“…Learning to switch. Since there are no actual labels for S to indicate whether it is correct to ask human or not, some previous work (Woodward and Finn, 2016;Rajendran et al, 2019) proposes to use the REINFORCE algorithm (Williams, 1992) for weakly-supervised training, but their reward settings fail to penalize the case when the model asks human while it can give right prediction, which may lead to redundant requests. To consider this effect, we propose a new reward definition here.…”
Section: Model Architecturementioning
confidence: 99%
See 3 more Smart Citations
“…Learning to switch. Since there are no actual labels for S to indicate whether it is correct to ask human or not, some previous work (Woodward and Finn, 2016;Rajendran et al, 2019) proposes to use the REINFORCE algorithm (Williams, 1992) for weakly-supervised training, but their reward settings fail to penalize the case when the model asks human while it can give right prediction, which may lead to redundant requests. To consider this effect, we propose a new reward definition here.…”
Section: Model Architecturementioning
confidence: 99%
“…Rajendran et al (2018) learn dialogs with multiple possible answers. Our work is inspired by the work of (Rajendran et al, 2019; propose to solve unseen user behaviors through human-machine teamwork. The research of Chen et al, 2017;Lu et al, 2019) also show the advantages of incorporating the role of human to teach online.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…However, due to the complexity of human conversation, auto-chatbot can hardly meet all users' needs, while its potential failure perceives skepticism. AI-enabled customer service, for instance, may trigger unexpected business losses because of chatbot failures (Radziwill and Benton, 2017;Rajendran et al, 2019). Moreover, for chatbot adoption in sensitive areas, such as healthcare (Chung and Park, 2019) and criminal justice (Wang et al, 2020a), any subtle statistical miscalculation may trigger serious health and legal Figure 1: A snippet of a moderately satisfied customer service dialogue.…”
Section: Introductionmentioning
confidence: 99%