Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval 2021
DOI: 10.1145/3404835.3463238
|View full text |Cite
|
Sign up to set email alerts
|

Pyserini: A Python Toolkit for Reproducible Information Retrieval Research with Sparse and Dense Representations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
93
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
5

Relationship

1
9

Authors

Journals

citations
Cited by 194 publications
(93 citation statements)
references
References 21 publications
0
93
0
Order By: Relevance
“…Accordingly, we evaluate the system runs with The New York Times Annotated Corpus and the topics of TREC Common Core 2017 [1]. As part of our experiments, we exploit the interactive search possibilities of the Pyserini toolkit [29]. We index the Core17 test collection with the help of Anserini [44] and the default indexing options as provided in the regression guide 4 .…”
Section: Datasets and Implementation Detailsmentioning
confidence: 99%
“…Accordingly, we evaluate the system runs with The New York Times Annotated Corpus and the topics of TREC Common Core 2017 [1]. As part of our experiments, we exploit the interactive search possibilities of the Pyserini toolkit [29]. We index the Core17 test collection with the help of Anserini [44] and the default indexing options as provided in the regression guide 4 .…”
Section: Datasets and Implementation Detailsmentioning
confidence: 99%
“…We calculated the TF-IDF index using DrQA implementation for all unigrams and bigrams with 2 24 buckets. Inspired by the criticism of choosing weak baselines presented in [38], we decided to validate our TF-IDF baseline against the proposed Anserini toolkit implemented by Pyserini [39].…”
Section: Document Retrievalmentioning
confidence: 99%
“…For the dense retrievers used in SPAR, we directly take the publicly released checkpoints without retraining to combine with Λ. We use Pyserini (Lin et al, 2021a) for all sparse models used in this work including BM25 and UniCOIL.…”
Section: Implementation Detailsmentioning
confidence: 99%