Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2022
DOI: 10.18653/v1/2022.acl-long.274
|View full text |Cite
|
Sign up to set email alerts
|

Automatic Error Analysis for Document-level Information Extraction

Abstract: Document-level information extraction (IE)tasks have recently begun to be revisited in earnest using the end-to-end neural network techniques that have been successful on their sentence-level IE counterparts. Evaluation of the approaches, however, has been limited in a number of dimensions. In particular, the precision/recall/F1 scores typically reported provide few insights on the range of errors the models make. We build on the work of Kummerfeld and Klein (2013) to propose a transformation-based framework f… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 23 publications
0
4
0
Order By: Relevance
“…In recent years, it has become standard to evaluate only on the string-fill slots, plus the template type (Chambers and Jurafsky, 2011;Du et al, 2021b;Das et al, 2022;Chen et al, 2023). Du et al (2021b) thus proposed a version of Eq.…”
Section: Muc-4: Recent Workmentioning
confidence: 99%
“…In recent years, it has become standard to evaluate only on the string-fill slots, plus the template type (Chambers and Jurafsky, 2011;Du et al, 2021b;Das et al, 2022;Chen et al, 2023). Du et al (2021b) thus proposed a version of Eq.…”
Section: Muc-4: Recent Workmentioning
confidence: 99%
“…In NLP, there are some task-specific works on automatic error analysis such as Das et al (2022) on document-level information extraction, Kummerfeld and Klein (2013) on coreference resolution, Popović and Ney (2011) on machine translation, and etc. There is also extensive research conducted on different model evaluations to see whether models make erroneous datapoints in certain types of noising datapoints (Belinkov and Bisk, 2017;Rychalska et al, 2019) or adversarial datapoints (Ribeiro et al, 2018;Iyyer et al, 2018).…”
Section: Related Workmentioning
confidence: 99%
“…Nevertheless, many IE tasks require a more comprehensive understanding that often extends to the entire input document, leading to challenges such as length and multiple events when embedding full documents. Consequently, document-level datasets continue to pose challenges for even the most advanced models today (Das et al, 2022).…”
Section: Introductionmentioning
confidence: 99%