Proceedings of the 28th International Conference on Computational Linguistics 2020
DOI: 10.18653/v1/2020.coling-main.135
|View full text |Cite
|
Sign up to set email alerts
|

KnowDis: Knowledge Enhanced Data Augmentation for Event Causality Detection via Distant Supervision

Abstract: Modern models of event causality detection (ECD) are mainly based on supervised learning from small hand-labeled corpora. However, hand-labeled training data is expensive to produce, low coverage of causal expressions and limited in size, which makes supervised methods hard to detect causal relations between events. To solve this data lacking problem, we investigate a data augmentation framework for ECD, dubbed as Knowledge Enhanced Distant Data Augmentation (KnowDis). Experimental results on two benchmark dat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
33
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

3
5

Authors

Journals

citations
Cited by 40 publications
(33 citation statements)
references
References 21 publications
0
33
0
Order By: Relevance
“…Caselli and Vossen (2017) (Chaudhary, 2020). Zuo et al (2020) solved the data lacking problem of ECI with the distantly supervised labeled training data. However, including the distant supervision, most of the existing data augmentation methods for NLP tasks are task-independent frameworks (Related work of data augmentation and dual learning are detailed in Appendix B).…”
Section: Related Workmentioning
confidence: 99%
“…Caselli and Vossen (2017) (Chaudhary, 2020). Zuo et al (2020) solved the data lacking problem of ECI with the distantly supervised labeled training data. However, including the distant supervision, most of the existing data augmentation methods for NLP tasks are task-independent frameworks (Related work of data augmentation and dual learning are detailed in Appendix B).…”
Section: Related Workmentioning
confidence: 99%
“…propose a mention masking generalization method and also consider the external structural knowledge. The very recent work (Zuo et al, 2020) propose a data augmentation method to alleviate the data lacking problem for the task. Regarding datasets construction, Mirza (2014) annotates the Causal-TimeBank dataset about event causal relations in the TempEval-3 corpus.…”
Section: Related Workmentioning
confidence: 99%
“…Our context encoder is based on the Transformer architecture (Vaswani et al, 2017). We adopt the BERT (Devlin et al, 2019) to encode the input sentence, 2 which has achieved the state-of-the-art performance for ECI task Zuo et al, 2020). After using BERT encoder to compute the contextual representations of the entire sentence, we concatenate representations of [CLS], e 1 and e 2 as the context representation regarding to the event pair (e 1 , e 2 ), namely…”
Section: Context Encodingmentioning
confidence: 99%
See 1 more Smart Citation
“…It will still suffer from data limitations and have no capacity to handling unseen contexts. Moreover, Zuo et al (2020) improved the performance of ECI with the distantly supervised labeled training data. However, their models are still limited to the unsatisfied qualities of the automatically generated data.…”
Section: Billy Finds His Childhood Teddy Bear >Causes/enables>mentioning
confidence: 99%