2017
DOI: 10.1007/s10791-017-9312-z
|View full text |Cite
|
Sign up to set email alerts
|

Statistical biases in Information Retrieval metrics for recommender systems

Abstract: There is an increasing consensus in the Recommender Systems community that the dominant error-based evaluation metrics are insufficient, and mostly inadequate, to properly assess the practical effectiveness of recommendations. Seeking to evaluate recommendation rankings—which largely determine the effective accuracy in matching user needs—rather than predicted rating values, Information Retrieval metrics have started to be applied for the evaluation of recommender systems. In this paper we analyse the main iss… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
91
0
3

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
1
1

Relationship

1
5

Authors

Journals

citations
Cited by 144 publications
(96 citation statements)
references
References 31 publications
2
91
0
3
Order By: Relevance
“…Φ D i cannot exist. On the other hand, previous studies [1,2,20] have emphasized the importance of modeling implicit feedback as part of the recommendation model. In the following section, we leverage the explicit model described in Equation (4) as a weakly supervised signal and augment it with the implicit feedback at the level of data pre-processing.…”
Section: Explicit Recommendation Modelmentioning
confidence: 99%
See 4 more Smart Citations
“…Φ D i cannot exist. On the other hand, previous studies [1,2,20] have emphasized the importance of modeling implicit feedback as part of the recommendation model. In the following section, we leverage the explicit model described in Equation (4) as a weakly supervised signal and augment it with the implicit feedback at the level of data pre-processing.…”
Section: Explicit Recommendation Modelmentioning
confidence: 99%
“…The task of rating prediction is measured by comparing the predicted ratings with the ground truth; metrics such as Root Mean Square Error (RMSE) are useful in this context. On the other hand, a ranking task is usually measured based on Information Retrieval (IR) metrics including Mean Reciprocal Rank (MRR) and normalized Discounted Cumulative Gain (nDCG) [1,2]. We evaluate the performance of the proposed model on both the rating prediction and ranking tasks.…”
Section: Datasets and Evaluation Metricsmentioning
confidence: 99%
See 3 more Smart Citations