2016
DOI: 10.1007/s11257-016-9174-x
|View full text |Cite
|
Sign up to set email alerts
|

Towards reproducibility in recommender-systems research

Abstract: Numerous recommendation approaches are in use today. However, comparing their effectiveness is a challenging task because evaluation results are rarely reproducible. In this article, we examine the challenge of reproducibility in recommender-system research. We conduct experiments using Plista's news recommender system, and Docear's research-paper recommender system. The experiments show that there are large discrepancies in the effectiveness of identical recommendation approaches in only slightly different sc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
41
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
6
2

Relationship

3
5

Authors

Journals

citations
Cited by 61 publications
(41 citation statements)
references
References 81 publications
(103 reference statements)
0
41
0
Order By: Relevance
“…• A working version of the source code is available or the code only has to be modified in minimal ways to work correctly. 3 • At least one dataset used in the original paper is available. A further requirement here is that either the originally-used train-test splits are publicly available or that they can be reconstructed based on the information in the paper.…”
Section: Research Methods 21 Collecting Reproducible Papersmentioning
confidence: 99%
See 1 more Smart Citation
“…• A working version of the source code is available or the code only has to be modified in minimal ways to work correctly. 3 • At least one dataset used in the original paper is available. A further requirement here is that either the originally-used train-test splits are publicly available or that they can be reconstructed based on the information in the paper.…”
Section: Research Methods 21 Collecting Reproducible Papersmentioning
confidence: 99%
“…Precisely speaking, we used a mix of replication and reproduction[12,35], i.e., we used both artifacts provided by the authors and our own artifacts. For the sake of readability, we will only use the term "reproducibility" in this paper 3. We did not apply modifications to the core algorithms.…”
mentioning
confidence: 99%
“…In the long-run, we hope to provide a platform to the information retrieval, digital library, and recommender systems community that helps conducting more reproducible and robust research in real-world scenarios [34,35]. To achieve this, we plan to add more partners on both sidesplatform partners who provide access to real users, and research partners who evaluate their novel algorithms via the living lab.…”
Section: Future Workmentioning
confidence: 99%
“…Unfortunately, university-based researchers struggle unless they closely collaborate with industry (e.g., [7]) or develop their own infrastructure and user base (e.g., [1]). Without online testing opportunities open to the research communities, they cannot employ online evaluation on a larger scale, which is the de-facto standard evaluation methodology in industry.…”
Section: Research Challengesmentioning
confidence: 99%