Proceedings of the 28th ACM International Conference on Information and Knowledge Management 2019
DOI: 10.1145/3357384.3358145
|View full text |Cite
|
Sign up to set email alerts
|

Modeling Long-Range Context for Concurrent Dialogue Acts Recognition

Abstract: In dialogues, an utterance is a chain of consecutive sentences produced by one speaker which ranges from a short sentence to a thousand-word post. When studying dialogues at the utterance level, it is not uncommon that an utterance would serve multiple functions. For instance, "Thank you. It works great. " expresses both gratitude and positive feedback in the same utterance. Multiple dialogue acts (DA) for one utterance breeds complex dependencies across dialogue turns. Therefore, DA recognition challenges a m… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(15 citation statements)
references
References 14 publications
(26 reference statements)
0
15
0
Order By: Relevance
“…Source distribution and major speakers statistics of the dataset are shown in Figures 3a and 3b, respectively. Since DAC and ER tasks are known to exploit the contextual features, i.e., dialogue history (Yu et al, 2019) so, utterances in the dataset are accompanied with their corresponding contextual utterances, which are typically preceding dialogue turns by the speakers participating in the dialogue. Each of the utterances contains three modalities: video, audio, and text.…”
Section: Emotion-da Dataset: Emotydamentioning
confidence: 99%
See 1 more Smart Citation
“…Source distribution and major speakers statistics of the dataset are shown in Figures 3a and 3b, respectively. Since DAC and ER tasks are known to exploit the contextual features, i.e., dialogue history (Yu et al, 2019) so, utterances in the dataset are accompanied with their corresponding contextual utterances, which are typically preceding dialogue turns by the speakers participating in the dialogue. Each of the utterances contains three modalities: video, audio, and text.…”
Section: Emotion-da Dataset: Emotydamentioning
confidence: 99%
“…A contextual self-attention system fused with hierarchical recurrent units was proposed by the authors of (Raheja and Tetreault, 2019) to develop a sequence label classifier. The authors of (Yu et al, 2019) proposed a method for the capture of long-range interactions that span a series of words using a Convolutional Network based approach. In (Saha et al, 2019), authors proposed several ML and DL based approaches such as Conditional Random Fields, clustering and word embeddings to identify DAs.…”
Section: Related Workmentioning
confidence: 99%
“…In our experiments, we split training/validation/testing datasets following Yu et al (2019) for MSDialog and Lee and Dernoncourt (2016) for MRDA. For two datasets, we first strip punctuation, and then we convert the characters into lower-case and tokenize the texts with NLTK 1 .…”
Section: Implementation Detailsmentioning
confidence: 99%
“…4) GA-Seq (Colombo et al, 2020) leverages a sequence to sequence approach to improving the modeling of tag sequentiality. 5) CRNN (Yu et al, 2019) is an adapted Convolutional Recurrent Neural Network (CRNN) that models the interactions between utterances of long-range context. 6) Di-alogueGCN (Ghosal et al, 2019) proposes a graph-based model that leverages self and inter-speaker interaction of the interlocutors to model conversational context for emotion recognition.…”
Section: Baselinesmentioning
confidence: 99%
See 1 more Smart Citation