Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021
DOI: 10.18653/v1/2021.emnlp-main.437
|View full text |Cite
|
Sign up to set email alerts
|

Learning from Noisy Labels for Entity-Centric Information Extraction

Abstract: Recent information extraction approaches have relied on training deep neural models. However, such models can easily overfit noisy labels and suffer from performance degradation. While it is very costly to filter noisy labels in large learning resources, recent studies show that such labels take more training steps to be memorized and are more frequently forgotten than clean labels, therefore are identifiable in training. Motivated by such properties, we propose a simple co-regularization framework for entity-… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 21 publications
(12 citation statements)
references
References 32 publications
0
12
0
Order By: Relevance
“…For future work, we plan to explore other strategies to better leverage the entity type information for RE with NLI and investigate if document-level NLI is also more challenging than sentence-level NLI. Another potential direction is to experiment with other DS techniques, such as integrating a denoising module to the RE model (Xiao et al, 2020) or using DS-trained models as a DS filter (Zhou and Chen, 2021).…”
Section: Discussionmentioning
confidence: 99%
“…For future work, we plan to explore other strategies to better leverage the entity type information for RE with NLI and investigate if document-level NLI is also more challenging than sentence-level NLI. Another potential direction is to experiment with other DS techniques, such as integrating a denoising module to the RE model (Xiao et al, 2020) or using DS-trained models as a DS filter (Zhou and Chen, 2021).…”
Section: Discussionmentioning
confidence: 99%
“…We will introduce methodologies that enhance the robustness of learning systems for IE in both their learning and inference phases. Those methodologies involve self-supervised denoising techniques for training noise-robust IE models based on coregularized knowledge distillation (Zhou and Chen, 2021;Liang et al, 2021), label re-weighting (Wang et al, 2019b) and label smoothing (Lukasik et al, 2020). Besides, we will also discuss about unsuper-vised techniques for out-of-distribution (OOD) detection (Zhou et al, 2021b;Hendrycks et al, 2020), prediction with abstention (Dhamija et al, 2018;Hendrycks et al, 2018) and novelty class detection (Perera and Patel, 2019) that seek to help the IE model identify invalid inputs or inputs with semantic shifts during its inference phase.…”
Section: Robust Learning and Inference For Ie [35min]mentioning
confidence: 99%
“…Following the works on Named Entity Recognition problem (Akbik et al, 2018;Yamada et al, 2020;Zhou and Chen, 2021), we used micro-average F1 score as our main measure for the Extraction Module. In addition for this module we added a detailed measure for each tag type i.e.…”
Section: Metricsmentioning
confidence: 99%