2017
DOI: 10.1609/aaai.v31i1.10961
|View full text |Cite
|
Sign up to set email alerts
|

Coherent Dialogue with Attention-Based Language Models

Abstract: We model coherent conversation continuation via RNN-based dialogue models equipped with a dynamic attention mechanism. Our attention-RNN language model dynamically increases the scope of attention on the history as the conversation continues, as opposed to standard attention (or alignment) models with a fixed input scope in a sequence-to-sequence model. This allows each generated word to be associated with the most relevant words in its corresponding conversation history. We evaluate the model on two popular d… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 33 publications
(9 citation statements)
references
References 25 publications
0
8
0
Order By: Relevance
“…Recent developments in deep learning have led to end-to-end approaches to dialogue using supervised learning, such as sequence-to-sequence models (Dušek and Jurcicek, 2016;Eric and Manning, 2017), hierarchical models (Serban et al, 2017), attention (Mei et al, 2017;Chen et al, 2019), andTransformer-based models (Wu et al, 2021;Hosseini-Asl et al, 2020;Peng et al, 2020;Adiwardana et al, 2020). However, supervised learning only allows an agent to imitate behaviors, requires optimal data, and does not allow agents to exceed human performance.…”
Section: Related Workmentioning
confidence: 99%
“…Recent developments in deep learning have led to end-to-end approaches to dialogue using supervised learning, such as sequence-to-sequence models (Dušek and Jurcicek, 2016;Eric and Manning, 2017), hierarchical models (Serban et al, 2017), attention (Mei et al, 2017;Chen et al, 2019), andTransformer-based models (Wu et al, 2021;Hosseini-Asl et al, 2020;Peng et al, 2020;Adiwardana et al, 2020). However, supervised learning only allows an agent to imitate behaviors, requires optimal data, and does not allow agents to exceed human performance.…”
Section: Related Workmentioning
confidence: 99%
“…Constrained recurrent models are also used to generate online product reviews of certain topic, sentiment, style and length [28], affective dialogue responses [40], or for modeling participant roles and topics in conversational systems [102].…”
Section: E Adapting Existing Models and Architectures To Accommodate ...mentioning
confidence: 99%
“…Recently, short text conversation has been popular. The system receives a short dialog context and generates a response using statistical machine translation or seq-to-seq networks (Ritter, Cherry, and Dolan 2011;Vinyals and Le 2015;Shang, Lu, and Li 2015;Serban et al 2016;Li et al 2016;Mei, Bansal, and Walter 2017). In contrast to response generation, the retrieval-based approach uses a ranking model to select the highest scoring response from candidates (Lu and Li 2013;Hu et al 2014;Ji, Lu, and Li 2014;Wang et al 2015).…”
Section: Related Workmentioning
confidence: 99%