Proceedings of the 18th BioNLP Workshop and Shared Task 2019
DOI: 10.18653/v1/w19-5031
|View full text |Cite
|
Sign up to set email alerts
|

Transfer Learning for Causal Sentence Detection

Abstract: We consider the task of detecting sentences that express causality, as a step towards mining causal relations from texts. To bypass the scarcity of causal instances in relation extraction datasets, we exploit transfer learning, namely ELMO and BERT, using a bidirectional GRU with self-attention (BIGRUATT) as a baseline. We experiment with both generic public relation extraction datasets and a new biomedical causal sentence detection dataset, a subset of which we make publicly available. We find that transfer l… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 25 publications
(17 citation statements)
references
References 26 publications
0
16
1
Order By: Relevance
“…The studies of [51,56,89,91,98] employ different deep learning models in order to extract causality from the SemEval-2010 Task 8 dataset. Xu et al [91] use LSTM to learn higherlevel semantic and syntactic representations along the shortest dependency path (SDP), while Li et al [56] combine BiLSTM with multi-head self-attention (MHSA) to direct attention to long-range dependencies between words.…”
Section: Explicit Intra-sentential Causalitymentioning
confidence: 99%
See 2 more Smart Citations
“…The studies of [51,56,89,91,98] employ different deep learning models in order to extract causality from the SemEval-2010 Task 8 dataset. Xu et al [91] use LSTM to learn higherlevel semantic and syntactic representations along the shortest dependency path (SDP), while Li et al [56] combine BiLSTM with multi-head self-attention (MHSA) to direct attention to long-range dependencies between words.…”
Section: Explicit Intra-sentential Causalitymentioning
confidence: 99%
“…This is the main reason for the F-score of [98] to be lower than [89] by 3.2%. Kyriakakis et al [51] explore the application of PTMs like BERT and ELMO [72] in the context of CE, by using bidirectional GRU with self-attention (BIGRUATT) as the baseline. The experimental results show that PTMs are only helpful for datasets with hundreds of training examples, and that BIGRUATT reaches a performance plateau when thousands of training instances are available.…”
Section: Explicit Intra-sentential Causalitymentioning
confidence: 99%
See 1 more Smart Citation
“…4. Opposed to pattern-matching are feature-based classification methods: recent papers apply neural networks and exploit-similarly to our approach-the Transfer Learning capability of BERT [31]. However, we see a number of problems with these papers regarding the realization of our described RE use cases: First, neither the code nor a demonstration is published, making it difficult to reproduce the results and test the performance on data from the RE domain.…”
Section: Causality In Requirements Engineeringmentioning
confidence: 99%
“…Hence, we believe systems to detect and parse causal relationships would perform well on this task as well. With respect to the first subtask, fine-tuning neural models using information-rich word embeddings seems to form the state-of-the-art (Kyriakakis et al, 2019) and forms the backbone of our submitted approach as well. The second subtask bears a resemblance to relation-entity extraction (of which cause-effect may be considered a specific case), which has also been recently dominated by neural models (Li and Tian, 2020;Soares et al, 2019).…”
Section: Previous and Related Workmentioning
confidence: 99%