2023
DOI: 10.1016/j.csl.2023.101482
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating factual accuracy in complex data-to-text

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 19 publications
(24 citation statements)
references
References 31 publications
0
5
0
Order By: Relevance
“…Thomson and Reiter (2021) noted that even within a single criterion, two annotators could provide sets of errors that only partially overlap, yet can both be considered valid representations of the same complex underlying problem. In addition to calculating agreement, annotators could check each other's annotations and indicate whether they consider them one valid way of describing the underlying problems Thomson et al (2023).…”
Section: Error Taxonomies and Standardizationmentioning
confidence: 99%
“…Thomson and Reiter (2021) noted that even within a single criterion, two annotators could provide sets of errors that only partially overlap, yet can both be considered valid representations of the same complex underlying problem. In addition to calculating agreement, annotators could check each other's annotations and indicate whether they consider them one valid way of describing the underlying problems Thomson et al (2023).…”
Section: Error Taxonomies and Standardizationmentioning
confidence: 99%
“…They should be provided with the annotation guidelines ( §4.5), and then be asked to annotate texts where the errors are known (but not visible). The solutions would ideally be created by experts, although in some cases, solutions created by researchers may be sufficient (Thomson and Reiter, 2020). It should be decided in advance what the threshold is to accept annotators for the reaming work, and, if they fail, whether to provide additional training or find other candidates.…”
Section: Number Of Annotatorsmentioning
confidence: 99%
“…For example, Costa et al (2015) describe how different taxonomies of errors in Machine Translation build on each other. In NLG, if you are working on data-to-text, then you could take Thomson and Reiter's (2020) taxonomy as a starting point. Alternatively, Dou et al (2021) present a crowd-sourced error annotation called SCARECROW.…”
Section: Bottom-up Approaches First Identify Differentmentioning
confidence: 99%
See 2 more Smart Citations