2022
DOI: 10.1007/978-3-030-99736-6_24
|View full text |Cite
|
Sign up to set email alerts
|

HC4: A New Suite of Test Collections for Ad Hoc CLIR

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
6
2

Relationship

1
7

Authors

Journals

citations
Cited by 18 publications
(10 citation statements)
references
References 31 publications
0
10
0
Order By: Relevance
“…The HC4 [14] dataset was employed as a validation set for the selection of the most optimal translators and first-stage retrievers due to the shared language coverage between it and NeuCLIR, as well as the existence of overlapping annotated query-document pairs between the two datasets. The RRF and SPLADE first-stage runs were provided by the NLE and h2loo teams, however, at the time of submission, the NLE team did not have a SPLADE model available for Chinese.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…The HC4 [14] dataset was employed as a validation set for the selection of the most optimal translators and first-stage retrievers due to the shared language coverage between it and NeuCLIR, as well as the existence of overlapping annotated query-document pairs between the two datasets. The RRF and SPLADE first-stage runs were provided by the NLE and h2loo teams, however, at the time of submission, the NLE team did not have a SPLADE model available for Chinese.…”
Section: Methodsmentioning
confidence: 99%
“…Regarding the branch of multilingual and CLIR, it is crucial to have access to appropriate datasets that can be used for both development and evaluation of models. In recent years, several datasets that support research in this area have been made publicly available, such as Fire [22,21], MLQA [15], NTCIR [30], Mr. Tydi [40], and HC4 [13].…”
Section: Related Workmentioning
confidence: 99%
“…We perform relevance assessment on a graded scale (between 0 and 3) using developed guidelines to ensure a consistent assessment process. Guidelines take inspiration from those of HC4 [19] and are adapted for our tasks (full guidelines online).…”
Section: Relevance Criteriamentioning
confidence: 99%
“…We evaluate the final retrieval models on HC4 [26], a newly constructed evaluation collection for CLIR, for Chinese and Persian, NTCIR [31] for Chinese, CLEF 08-09 for Persian [1,14], and CLEF 03 [4] for French and German. HC4 consists of 50 topics for each language.…”
Section: Datasetsmentioning
confidence: 99%