Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing 2017
DOI: 10.18653/v1/d17-1231
|View full text |Cite
|
Sign up to set email alerts
|

Using Context Information for Dialog Act Classification in DNN Framework

Abstract: Previous work on dialog act (DA) classification has investigated different methods, such as hidden Markov models, maximum entropy, conditional random fields, graphical models, and support vector machines. A few recent studies explored using deep learning neural networks for DA classification, however, it is not clear yet what is the best method for using dialog context or DA sequential information, and how much gain it brings. This paper proposes several ways of using context information for DA classification,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

10
146
0
9

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 87 publications
(165 citation statements)
references
References 11 publications
10
146
0
9
Order By: Relevance
“…To try and capture interactions between speakers, Kalchbrenner and Blunsom [10] used a Hierarchical CNN sentence model in conjunction with a RNN discourse model and condition the recurrent and output weights on the current speaker. Liu et al, [16] examine several different CNN and LSTM based architectures that incorporate different context information such as speaker change and dialogue history. Their work shows that including context information consistently yielded improvements over their baseline system.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…To try and capture interactions between speakers, Kalchbrenner and Blunsom [10] used a Hierarchical CNN sentence model in conjunction with a RNN discourse model and condition the recurrent and output weights on the current speaker. Liu et al, [16] examine several different CNN and LSTM based architectures that incorporate different context information such as speaker change and dialogue history. Their work shows that including context information consistently yielded improvements over their baseline system.…”
Section: Related Workmentioning
confidence: 99%
“…This view has led much of the neural network based research to model the semantic content of a sentence, in conjunction with some other contextual information, such as previous utterance or DA sequences, or a change in speaker turn, to predict the appropriate DA for the current utterance [10,14]. Including such historical and contextual information has been shown to improve classification accuracy [16], and likely must be considered for any sophisticated classification model. However, motivated to examine the importance of different lexical and syntactic features, and their contribution towards DA classification, in this work each utterance is considered individually and out-of-context, that is, without any other contextual or historical information.…”
Section: Introductionmentioning
confidence: 99%
“…For modeling intents of multi-turn dialogue dataset, SwDA, we use hierarchical CNN (CNN+CNN) similar to the structure described in [23]. We illustrate the experiment flowchart in Figure 3.…”
Section: Cnn+cnnmentioning
confidence: 99%
“…However, there is no standard scheme for splitting SwDA dataset into training, validation and test sets. We follow the data splitting scheme suggested in [23] by randomly sampling 80% of conversation for the training set, 10% for the validation set and 10% for the test set. There are 162862 utterances for training, 20784 utterances for validation and 20146 utterances for testing.…”
Section: Datasetsmentioning
confidence: 99%
“…Our work represents short user utterances using recurrent neural networks, and additionally models dialogue context using a hierarchical recurrent neural network. Such dialogue-level models have also been proposed in [22] for dialogue act tagging of human-human social phone conversations. Previous studies mainly considered DA tagging of multi-human conversations, such as the Switchboard [15] corpus and meetings, such as the ICSI meeting corpus [23], whereas our focus lies on modeling system-side DAs.…”
Section: Related Workmentioning
confidence: 99%