Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing 2016
DOI: 10.18653/v1/d16-1127
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning for Dialogue Generation

Abstract: Recent neural models of dialogue generation offer great promise for generating responses for conversational agents, but tend to be shortsighted, predicting utterances one at a time while ignoring their influence on future outcomes. Modeling the future direction of a dialogue is crucial to generating coherent, interesting dialogues, a need which led traditional NLP models of dialogue to draw on reinforcement learning. In this paper, we show how to integrate these goals, applying deep reinforcement learning to m… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

4
692
2
3

Year Published

2017
2017
2022
2022

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 912 publications
(726 citation statements)
references
References 39 publications
4
692
2
3
Order By: Relevance
“…For task-specific dialogue Zhao and Eskenazi, 2016;Cuayáhuitl et al, 2016;Williams and Zweig, 2016;Li et al, 2017b,c;Peng et al, 2017), the reward function is usually based on task completion rate, and thus is easy to define. For the much harder problem of open-domain dialogue generation (Li et al, 2016e;Weston, 2016), hand-crafted reward functions are used to capture desirable conversation properties. Li et al (2016d) propose DRL-based diversitypromoting Beam Search (Koehn et al, 2003) for response generation.…”
Section: Related Work and Contributionsmentioning
confidence: 99%
“…For task-specific dialogue Zhao and Eskenazi, 2016;Cuayáhuitl et al, 2016;Williams and Zweig, 2016;Li et al, 2017b,c;Peng et al, 2017), the reward function is usually based on task completion rate, and thus is easy to define. For the much harder problem of open-domain dialogue generation (Li et al, 2016e;Weston, 2016), hand-crafted reward functions are used to capture desirable conversation properties. Li et al (2016d) propose DRL-based diversitypromoting Beam Search (Koehn et al, 2003) for response generation.…”
Section: Related Work and Contributionsmentioning
confidence: 99%
“…Recent years have seen neural networks being applied to all key parts of the typical modern IR pipeline, such core ranking algorithms [26,42,51], click models [9,10], knowledge graphs [8,35], text similarity [28,47], entity retrieval [52,53], language modeling [5], question answering [22,56], and dialogue systems [34,54].…”
Section: Motivationmentioning
confidence: 99%
“…Targeting this newly emerging demand, some models have been proposed to respond by generating natural language replies on the y, rather than by (re)ranking a xed set of items or extracting passages from existing pages. Examples are conversational and dialog systems [7,34,54] or machine reading and question answering tasks where the model either infers the answer from unstructured data, like textual documents that do not necessarily feature the answer literally [21,22,46,56], or generates natural language given structured data, like data from knowledge graphs or from external memories [1,18,33,37,40].…”
Section: Objectivesmentioning
confidence: 99%
“…Reinforcement Learning can be used to improve dialogue managers, e.g. for transitions between dialogue states (Rieser and Lemon 2011), for nongoal-orientated dialogues (Li et al 2016), for botbot dialogues and for inventing new languages by agents (Das et al 2017). …”
Section: Deep Reinforcement Learningmentioning
confidence: 99%