Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.636
|View full text |Cite
|
Sign up to set email alerts
|

Extracting Event Temporal Relations via Hyperbolic Geometry

Abstract: Detecting events and their evolution through time is a crucial task in natural language understanding. Recent neural approaches to event temporal relation extraction typically map events to embeddings in the Euclidean space and train a classifier to detect temporal relations between event pairs. However, embeddings in the Euclidean space cannot capture richer asymmetric relations such as event temporal relations. We thus propose to embed events into hyperbolic spaces, which are intrinsically oriented at modeli… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
13
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 17 publications
(13 citation statements)
references
References 32 publications
0
13
0
Order By: Relevance
“…Compared to ECONET and TGT, which use a larger pre-trained language model, or TGT and HGRU, which use networks with complex structures followed RoBERTa base or BERT Large, TempACL enables a smaller and simpler model which only contains a RoBERTa base and two fully connected layers to achieve the state-of-the-art performance. Method TB-Dense MATRES JCL RoBERTa base -78.8 ECONET RoBERTa Large 66.8 79.3 TGT BERT Large 66.7 80.3 Poincaré Event Embeddings (Tan et al, 2021) RoBERTa base -78.9 HGRU+knowledge (Tan et al, 2021) RoBERTa base -80.5 CERT (Fang et al, 2020) RoBERTa We observe that, TempACL make improvements of 5.37%F 1 and 1.81%F 1 on TB-Dense and MATRES respectively compared with the baseline model. In this section, we first qualitatively analyze key samples, and then we do the ablation experiments to further study the effects of patient strategies and label-aware contrastive learning loss.…”
Section: Resultsmentioning
confidence: 82%
See 2 more Smart Citations
“…Compared to ECONET and TGT, which use a larger pre-trained language model, or TGT and HGRU, which use networks with complex structures followed RoBERTa base or BERT Large, TempACL enables a smaller and simpler model which only contains a RoBERTa base and two fully connected layers to achieve the state-of-the-art performance. Method TB-Dense MATRES JCL RoBERTa base -78.8 ECONET RoBERTa Large 66.8 79.3 TGT BERT Large 66.7 80.3 Poincaré Event Embeddings (Tan et al, 2021) RoBERTa base -78.9 HGRU+knowledge (Tan et al, 2021) RoBERTa base -80.5 CERT (Fang et al, 2020) RoBERTa We observe that, TempACL make improvements of 5.37%F 1 and 1.81%F 1 on TB-Dense and MATRES respectively compared with the baseline model. In this section, we first qualitatively analyze key samples, and then we do the ablation experiments to further study the effects of patient strategies and label-aware contrastive learning loss.…”
Section: Resultsmentioning
confidence: 82%
“…We follow the official split strategy that uses TimeBank and AQUAINT for training and Platinum for testing. We also follow the previous works (Ning et al, 2019;Tan et al, 2021) that randomly select 20 percents of the official train documents as dev set.…”
Section: Experiments and Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Mathur et al [ 43 ] put forward a method leveraging rhetorical discourse features and temporal arguments from semantic role labels as well as local syntactic features through a Gated Relational-GCN model. Tan et al [ 44 ] tried to embed events into hyperbolic spaces and train a classifier to capture temporal relations. Breitfeller et al [ 45 ] developed a novel framework to explore semantic information provided by explicit textual time clues.…”
Section: Related Workmentioning
confidence: 99%
“…BEFORE and AFTER that follow the arrows denote the extracted TEMPREL's from the sentences by . proaches, for which recent studies also incorporate advanced learning and inference techniques such as structured prediction (Ning et al, 2017(Ning et al, , 2018bHan et al, 2019;Tan et al, 2021), graph representation (Mathur et al, 2021;Zhang et al, 2022), data augmentation (Ballesteros et al, 2020;Trong et al, 2022), and indirect supervision (Zhao et al, 2021;. These models are prevalently built upon pretrained language models (PLMs) and fine-tuned on a small set of annotated documents, e.g., TimeBank-Dense (Cassidy et al, 2014), MATRES (Ning et al, 2018c), and TDDiscourse (Naik et al, 2019).…”
Section: Introductionmentioning
confidence: 99%