Proceedings of the 30th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval 2007
DOI: 10.1145/1277741.1277879
|View full text |Cite
|
Sign up to set email alerts
|

Intra-assessor consistency in question answering

Abstract: In this paper we investigate the consistency of answer assessment in a complex question answering task examining features of assessor consistency, types of answers and question type.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
7
0

Year Published

2008
2008
2018
2018

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(8 citation statements)
references
References 3 publications
1
7
0
Order By: Relevance
“…While the idea of relevance being inherently subjective has been pointed out by many researchers (e.g., see references [29] and more recently [21]), we note that in community QA a large fraction of the questions are subjective, compounding the problem of both relevance assessment (which is no longer meaningful). Information seeker satisfaction has been studied in ad-hoc IR context in [11] (refer to [15] for an overview), but studies have been limited by lack of realistic user feedback on whole-result satisfaction and instead worked primarily within the Cranfield evaluation model.…”
Section: Related Workmentioning
confidence: 78%
See 1 more Smart Citation
“…While the idea of relevance being inherently subjective has been pointed out by many researchers (e.g., see references [29] and more recently [21]), we note that in community QA a large fraction of the questions are subjective, compounding the problem of both relevance assessment (which is no longer meaningful). Information seeker satisfaction has been studied in ad-hoc IR context in [11] (refer to [15] for an overview), but studies have been limited by lack of realistic user feedback on whole-result satisfaction and instead worked primarily within the Cranfield evaluation model.…”
Section: Related Workmentioning
confidence: 78%
“…This is in contrast to the more traditional relevance-based assessment that is often done by judges different from the original information seeker, which may result in ratings that do not agree with the target user. While the idea of relevance being inherently subjective has been pointed out in the past (e.g., see references [29] and more recently [21]), nowhere does the problem of subjective relevance arise more prominently than within Community QA, where many of the questions are inherently subjective, complex, ill-formed, or often all of the above. The problem of complex and subjective QA has only recently started to be addressed in the question answering community, most recently as the first opinion QA track in TREC [7].…”
Section: Introductionmentioning
confidence: 99%
“…While the idea of relevance being inherently subjective has been pointed out by many researchers (e.g., see Zobel [1998] and more recently Ruthven et al [2007]), we note that in community QA a large fraction of the questions are subjective, compounding the problem of both relevance assessment (which is no longer meaningful). Information seeker satisfaction has been studied in ad-hoc IR context in Harter and Hert [1997] (refer to Kobayashi and Takeda [2000] for an overview), but studies have been limited by lack of realistic user feedback on whole-result satisfaction and instead worked primarily within the Cranfield evaluation model.…”
Section: Related Workmentioning
confidence: 78%
“…This is in contrast to the more traditional relevance-based assessment that is often done by judges different from the original information seeker, which may result in ratings that do not agree with the target user. While the idea of relevance being inherently subjective has been pointed out in the past (e.g., see Zobel [1998] and more recently Ruthven et al [2007]), nowhere does the problem of subjective relevance arise more prominently than within Community QA, where many of the questions are inherently subjective, complex, ill-formed, or often all of the above. The problem of complex and subjective QA has only recently started to be addressed in the question answering community, most recently as the first opinion QA track in TREC [Dang et al 2007].…”
Section: Introductionmentioning
confidence: 99%
“…Second, the quality of answer lists shown in Figure 1 includes the primary assessor's judgment after the interaction. Ideally, the interactive assessor's judgment at the individual nugget level should be obtained at the same time that they interact with the side-by-side interface, since judgments may change as a topic becomes more and more familiar [3]. This is also borne out by a related experiment: as part of our participation in the ciQA task, we also used another interface to gather assessor judgments of individual answer strings.…”
Section: Discussionmentioning
confidence: 99%