Proceedings of the Fourth ACM International Conference on Web Search and Data Mining 2011
DOI: 10.1145/1935826.1935864
|View full text |Cite
|
Sign up to set email alerts
|

Using graded-relevance metrics for evaluating community QA answer selection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
26
0

Year Published

2011
2011
2023
2023

Publication Types

Select...
5
3
2

Relationship

1
9

Authors

Journals

citations
Cited by 40 publications
(28 citation statements)
references
References 13 publications
2
26
0
Order By: Relevance
“…Similar observations have been made in evaluating CQA answer ranking 24) . While we cannot conclude how many assessors are required for reliable experiments from our present study, it is possible that ensuring the quality of annotators is more important than the number of annotators 24) .…”
Section: Classification Results With Different Annotation Datasupporting
confidence: 77%
“…Similar observations have been made in evaluating CQA answer ranking 24) . While we cannot conclude how many assessors are required for reliable experiments from our present study, it is possible that ensuring the quality of annotators is more important than the number of annotators 24) .…”
Section: Classification Results With Different Annotation Datasupporting
confidence: 77%
“…Request permissions from permissions@acm.org. SIGIR'14, July [6][7][8][9][10][11]2014 level measures one would need to manually assess all possible passages, which is not practical. As an alternative, character-based measures have been developed that treat each character as a document and evaluate them using existing precision and recall measures [1,5].…”
Section: Introductionmentioning
confidence: 99%
“…Fichman found that the quality of answers was significantly improved only in terms of answer completeness and verifiability, rather than the answer accuracy. Sakai et al [13] proposed system evaluation methods for the task of selecting or ranking answers for a given question. They noticed that the asker-selected best answers might be biased and even if they were not, there might be other good answers besides the best ones.…”
Section: Retrieving High-quality Answers In Cqa Sitesmentioning
confidence: 99%