2009
DOI: 10.1007/978-3-642-04447-2_35
|View full text |Cite
|
Sign up to set email alerts
|

Overview of the Answer Validation Exercise 2008

Abstract: The Answer Validation Exercise at the Cross Language Evaluation Forum is aimed at developing systems able to decide whether the answer of a Question Answering system is correct or not. We present here the exercise description, the changes in the evaluation methodology with respect to the first edition, and the results of this second edition (AVE 2007). The changes in the evaluation methodology had two objectives: the first one was to quantify the gain in performance when more sophisticated validation modules a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2009
2009
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 30 publications
(11 citation statements)
references
References 15 publications
0
11
0
Order By: Relevance
“…The best result was obtained by DFKI in English domain with precision and recall were about 44% and 71%. Last report from AVE was done by [8] in AVE 2008. They presented the changes in the evaluation with respect to the last edition, and the results of this third edition.…”
Section: Fig2 Approaches Based On Av's Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…The best result was obtained by DFKI in English domain with precision and recall were about 44% and 71%. Last report from AVE was done by [8] in AVE 2008. They presented the changes in the evaluation with respect to the last edition, and the results of this third edition.…”
Section: Fig2 Approaches Based On Av's Methodsmentioning
confidence: 99%
“…The meetings were held three times in 2006-2008 [6][7][8] which proposed Recognition Textual Entailment (RTE) [9] as the primary approach where the hypotheses have been built semiautomatically turning the questions and the answers into an affirmative form. Participant systems must return a value YES or NO for each pair of text-hypothesis to indicate if the text entails the hypothesis or not.…”
Section: Introductionmentioning
confidence: 99%
“…Textual entailment has never been applied explicitly to fact checking problems, but they obviously meet at some points [37]. Many evaluation campaigns and benchmarks are related to textual entailment, as well as paraphrase detection in general, among which PASCAL challenge [13], Answer Validation Exercise [71], the MSRP paraphrase corpus [15] or the SNLI corpus [10]. Most of these tasks and data represent similarity between pairs of text as a binary yes/no classification decision.…”
Section: Claim Accuracy Assessmentmentioning
confidence: 99%
“…Section 3 presents the results analysis, and Section 4 exposes our conclusions and outlines some future work directions. Table 1 shows the answer validation results corresponding to our two submitted runs to Spanish AVE 2008 [3]. It also shows the results for the baseline (a 100% VALIDATED).…”
Section: Introductionmentioning
confidence: 98%