2023
DOI: 10.48550/arxiv.2301.06841
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Syntactically Robust Training on Partially-Observed Data for Open Information Extraction

Abstract: Open Information Extraction models have shown promising results with sufficient supervision. However, these models face a fundamental challenge that the syntactic distribution of training data is partially observable in comparison to the real world. In this paper, we propose a syntactically robust training framework that enables models to be trained on a syntactic-abundant distribution based on diverse paraphrase generation. To tackle the intrinsic problem of knowledge deformation of paraphrasing, two algorith… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(4 citation statements)
references
References 23 publications
0
4
0
Order By: Relevance
“…Despite the widespread interest in these benchmarks and the related OpenIE approaches provides promising results. However, the traditional peer-topeer matching-based evaluation can not measure the robustness of those approaches, where the syntax and expression may be various with underlying meaning (Qi et al, 2023). This work significantly fills the gap between traditional metrics and missed robustness evaluation for OpenIE and calls for more efforts in this research area.…”
Section: Related Workmentioning
confidence: 96%
See 3 more Smart Citations
“…Despite the widespread interest in these benchmarks and the related OpenIE approaches provides promising results. However, the traditional peer-topeer matching-based evaluation can not measure the robustness of those approaches, where the syntax and expression may be various with underlying meaning (Qi et al, 2023). This work significantly fills the gap between traditional metrics and missed robustness evaluation for OpenIE and calls for more efforts in this research area.…”
Section: Related Workmentioning
confidence: 96%
“…In order to analyze the syntactic divergence in the cliques, we need a metric to measure the syntactic correlation between two sentences. A fast and effective algorithm is the HWS distance proposed in (Qi et al, 2023), which calculates the syntactic tree distance between two sentences based on a hierarchically weighted matching strategy, where smaller weights imply a greater focus on the comparison of skeletons. The value domain of this is [0, 1], where 1 indicates the farthest distance.…”
Section: Syntactic Analysismentioning
confidence: 99%
See 2 more Smart Citations