2014
DOI: 10.1007/978-3-319-11382-1_22
|View full text |Cite
|
Sign up to set email alerts
|

Improving the Reproducibility of PAN’s Shared Tasks:

Abstract: Potthast, M.; Gollub, T.; Rangel, F.; Rosso, P.; Stamatatos, E.; Abstract This paper reports on the PAN 2014 evaluation lab which hosts three shared tasks on plagiarism detection, author identification, and author profiling.To improve the reproducibility of shared tasks in general, and PAN's tasks in particular, the Webis group developed a new web service called TIRA, which facilitates software submissions. Unlike many other labs, PAN asks participants to submit running softwares instead of their run output. … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
12
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
6
3
1

Relationship

2
8

Authors

Journals

citations
Cited by 28 publications
(12 citation statements)
references
References 38 publications
0
12
0
Order By: Relevance
“…For the submitted official results through TIRA (Potthast et al, 2014), we use the intersection model for all languages. Since we focus on the corpus selection, we do not perform additional preprocessing and we use the provided training datasets as they are.…”
Section: Resultsmentioning
confidence: 99%
“…For the submitted official results through TIRA (Potthast et al, 2014), we use the intersection model for all languages. Since we focus on the corpus selection, we do not perform additional preprocessing and we use the provided training datasets as they are.…”
Section: Resultsmentioning
confidence: 99%
“…Key goals of any empirical evaluation are to ensure a blind evaluation, its replicability, and its reproducibility. To facilitate these goals, we employed the cloud-based evaluation platform TIRA (Potthast et al, 2014), 13 which implements the evaluation as a service paradigm (Hanbury et al, 2015). In doing so, we depart from the traditional submission of system output to shared tasks, which lacks in these regards, toward the submission of working software.…”
Section: Evaluation Methodologymentioning
confidence: 99%
“…All of these test sets (Nivre et al, 2017b) were hidden from the participating teams until the shared task had ended. Using the TIRA environment (Potthast et al, 2014) provided for the shared task, participants could execute runs on them, but not see the outputs or the results.…”
Section: Test Splitsmentioning
confidence: 99%