Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval 2017
DOI: 10.1145/3077136.3080726
|View full text |Cite
|
Sign up to set email alerts
|

A Stream-based Resource for Multi-Dimensional Evaluation of Recommender Algorithms

Abstract: Recommender System research has evolved to focus on developing algorithms capable of high performance in online systems. is development calls for a new evaluation infrastructure that supports multi-dimensional evaluation of recommender systems. Today's researchers should analyze algorithms with respect to a variety of aspects including predictive performance and scalability. Researchers need to subject algorithms to realistic conditions in online A/B tests. We introduce two resources supporting such evaluation… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
2
2

Relationship

2
2

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 7 publications
0
3
0
Order By: Relevance
“…. NewsREEL allows for the evaluation of news recommender algorithms by either "replaying" a news recommendation situation using a static data set, or by providing access to a live data stream of user requests for news articles [Kille et al 2017]. The key component for the online setting that is introduced in Section 3.1.6 is the Open Recommender Platform (ORP) .…”
Section: Open Recommendation Platform (Orp)mentioning
confidence: 99%
“…. NewsREEL allows for the evaluation of news recommender algorithms by either "replaying" a news recommendation situation using a static data set, or by providing access to a live data stream of user requests for news articles [Kille et al 2017]. The key component for the online setting that is introduced in Section 3.1.6 is the Open Recommender Platform (ORP) .…”
Section: Open Recommendation Platform (Orp)mentioning
confidence: 99%
“…Existing studies performed offline evaluation and online evaluation separately (Gunawardana and Shani, 2009; Shani and Gunawardana, 2011; Garcin and Faltings, 2013; Freyne and Berkovsky, 2013; Gavalas et al , 2014; Hofmann et al , 2016; Kille et al , 2017). Offline evaluation recommends the most popular items, while this strategy is the poorest in an online evaluation; thus, our study is consistent with the literature (Garcin and Faltings, 2013).…”
Section: Discussionmentioning
confidence: 99%
“…Explicit attention to minimal necessary data will promote the development of such algorithms. We note that algorithms that use minimal personal data are useful to address news recommendation, where user IDs might be unstable or unavailable [16].…”
Section: Doing More With Lessmentioning
confidence: 99%