2018
DOI: 10.1016/j.neucom.2017.09.074
|View full text |Cite
|
Sign up to set email alerts
|

Learning better discourse representation for implicit discourse relation recognition via attention networks

Abstract: Humans comprehend the meanings and relations of discourses heavily relying on their semantic memory that encodes general knowledge about concepts and facts. Inspired by this, we propose a neural recognizer for implicit discourse relation analysis, which builds upon a semantic memory that stores knowledge in a distributed fashion. We refer to this recognizer as SeMDER. Starting from word embeddings of discourse arguments, SeMDER employs a shallow encoder to generate a distributed surface representation for a di… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 16 publications
(6 citation statements)
references
References 7 publications
0
6
0
Order By: Relevance
“…Injecting hierarchy information into prompts is also promising. For example, using top-level predictions to refine prompts of bottom levels can surpass soft prompts and hard prompts (Wang et al, 2022b). Nevertheless, how to employ LLMs to better involve hierarchy knowledge is still under investigation.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Injecting hierarchy information into prompts is also promising. For example, using top-level predictions to refine prompts of bottom levels can surpass soft prompts and hard prompts (Wang et al, 2022b). Nevertheless, how to employ LLMs to better involve hierarchy knowledge is still under investigation.…”
Section: Related Workmentioning
confidence: 99%
“…• GOLF (Jiang et al, 2022b): a global and local hierarchy-aware contrastive framework, to model and capture the information from these two kinds of hierarchies with the aid of contrastive learning.…”
Section: A3 Baseline Modelsmentioning
confidence: 99%
“…For the IDRR task, it is also argued that different words in arguments contribute differently in learning argument representations and interactions. Some researchers have proposed to augment conventional neural models with attention mechanisms, such as the attention-based CNNs [98,167] and attention-based RNNs [12,21,58,83]. For example, Zhang et al [167] proposed a full attention model that combines an inner attention model to process internal argument information and an outer attention model to exploit external world knowledge.…”
Section: Attention Mechanismmentioning
confidence: 99%
“…Some researchers have proposed to augment conventional neural models with attention mechanisms, such as the attention-based CNNs [98,167] and attention-based RNNs [12,21,58,83]. For example, Zhang et al [167] proposed a full attention model that combines an inner attention model to process internal argument information and an outer attention model to exploit external world knowledge. In the inner attention model, they employed the SCNN [165] to encode each argument as its original representation, which is then used to compute the attention weight of each word by a score function.…”
Section: Attention Mechanismmentioning
confidence: 99%
“…Based on bidirectional recurrent neural networks, Bahdanau et al [20] added the attention mechanism to the model to encode and decode the sentence in machine translation. Zhang et al [21] examined inner attention mechanism and outer attention mechanism in discourse representation for implicit discourse relation recognition. The result showed a marvelous improvement on marco- F 1 point is 1.61%.…”
Section: Related Workmentioning
confidence: 99%