Proceedings of the 24th Conference on Computational Natural Language Learning 2020
DOI: 10.18653/v1/2020.conll-1.28
|View full text |Cite
|
Sign up to set email alerts
|

In Media Res: A Corpus for Evaluating Named Entity Linking with Creative Works

Abstract: Annotation styles express guidelines that direct human annotators by explicitly stating the rules to follow when creating gold standard annotations of text corpora. These guidelines not only shape the gold standards they help create, but also influence the training and evaluation of Named Entity Linking (NEL) tools, since different annotation styles correspond to divergent views on the entities present in a document. Such divergence is particularly relevant for texts from the media domain containing references… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
1
1

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 24 publications
0
2
0
Order By: Relevance
“…Apart from that, they do not provide guidelines or suggestions that target the annotation process. Brasoveanu et al (2018) argue that an in-depth qualitative analysis of entity linking errors is necessary in order to efficiently improve entity linking systems. They categorize EL errors into five categories: knowledge base errors, dataset errors, annotator errors, NIL clustering errors and evaluation errors.…”
Section: Contributionsmentioning
confidence: 99%
“…Apart from that, they do not provide guidelines or suggestions that target the annotation process. Brasoveanu et al (2018) argue that an in-depth qualitative analysis of entity linking errors is necessary in order to efficiently improve entity linking systems. They categorize EL errors into five categories: knowledge base errors, dataset errors, annotator errors, NIL clustering errors and evaluation errors.…”
Section: Contributionsmentioning
confidence: 99%
“…and specific training of NER/NEL systems is needed for this case. In the ReTV project, a customisation of the Recogynze tool performed significantly better than the previously mentioned "off the shelf" systems [211]. These annotation tools are generically focused on the identification of the occurrence of entities in text, rather than the specific use case of describing the content of TV programming for personalisation services.…”
Section: Applications For Tv Content Annotationmentioning
confidence: 99%