Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1495
|View full text |Cite
|
Sign up to set email alerts
|

Event Representation Learning Enhanced with External Commonsense Knowledge

Abstract: Prior work has proposed effective methods to learn event representations that can capture syntactic and semantic information over text corpus, demonstrating their effectiveness for downstream tasks such as script event prediction. On the other hand, events extracted from raw texts lacks of commonsense knowledge, such as the intents and emotions of the event participants, which are useful for distinguishing event pairs when there are only subtle differences in their surface realizations. To address this issue, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
90
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 58 publications
(90 citation statements)
references
References 21 publications
0
90
0
Order By: Relevance
“…We propose two updates of the original HANs used in (Hu et al, 2018) and (Yang et al, 2016): inserting an event-layer between words and news levels, and representing events by pretrained contextualized language models such as BERT (Devlin et al, 2019) instead of bidirectional gated recurrent unit (bi-GRU) (Cho et al, 2014) plus attention networks. The encoding of financial information in the units of events is inspired by Ding et al (2014Ding et al ( , 2019. The difference is that, we replace the neural tensor network (Ding et al, 2014) by BERT.…”
Section: Bert-enhanced Hans With Three-level Attentionsmentioning
confidence: 99%
See 3 more Smart Citations
“…We propose two updates of the original HANs used in (Hu et al, 2018) and (Yang et al, 2016): inserting an event-layer between words and news levels, and representing events by pretrained contextualized language models such as BERT (Devlin et al, 2019) instead of bidirectional gated recurrent unit (bi-GRU) (Cho et al, 2014) plus attention networks. The encoding of financial information in the units of events is inspired by Ding et al (2014Ding et al ( , 2019. The difference is that, we replace the neural tensor network (Ding et al, 2014) by BERT.…”
Section: Bert-enhanced Hans With Three-level Attentionsmentioning
confidence: 99%
“…(2) too long news prevents the generalization ability of being embedded for financial information similarity computing. Generally, our proposed network can be seen as a combination of events (Ding et al, 2014(Ding et al, , 2019 represented by deep pretrained contextualized language models (Devlin et al, 2019) inside a hierarchical attention mechanisms (Hu et al, 2018;Yang et al, 2016). In Figure 2, we assume that there are N words in one financial event.…”
Section: Bert-enhanced Hans With Three-level Attentionsmentioning
confidence: 99%
See 2 more Smart Citations
“…Just to name a few, Trinh and Le (2018), He et al (2019) and Klein and Nabi (2019) use language models trained on huge text corpora to do inference on the WSC dataset. Ding et al (2019) use commonsense knowledge in Atomic (Sap et al, 2019a) and Event2mind (Rashkin et al, 2018) on downstream tasks such as script event prediction. Bi et al (2019) exploit external commonsense knowledge from ConceptNet (Speer et al, 2016)) in machine reading comprehension.…”
Section: Commonsense Datasetsmentioning
confidence: 99%