2004
DOI: 10.21236/ada460118
|View full text |Cite
|
Sign up to set email alerts
|

UMass at TREC 2004: Novelty and HARD

Abstract: For the TREC 2004 Novelty track, UMass participated in all four tasks. Although finding relevant sentences was harder this year than last, we continue to show marked improvements over the baseline of calling all sentences relevant, with a variant of tfidf being the most successful approach. We achieve 5-9% improvements over the baseline in locating novel sentences, primarily by looking at the similarity of a sentence to earlier sentences and focusing on named entities.For the High Accuracy Retrieval from Docum… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

3
221
0

Year Published

2009
2009
2016
2016

Publication Types

Select...
6
2
1

Relationship

2
7

Authors

Journals

citations
Cited by 191 publications
(224 citation statements)
references
References 33 publications
3
221
0
Order By: Relevance
“…This interpolated language model can then be used with Equation 4 to rank documents (Abdul-Jaleel et al, 2004). We will refer to this as the expanded query score of a document.…”
Section: Query Expansion With Word Embeddingsmentioning
confidence: 99%
“…This interpolated language model can then be used with Equation 4 to rank documents (Abdul-Jaleel et al, 2004). We will refer to this as the expanded query score of a document.…”
Section: Query Expansion With Word Embeddingsmentioning
confidence: 99%
“…In [14] they compared five methods to estimate the query language models: RM3 and RM4 [1]; a divergence minimization model (DMM) and a simple mixture model (SMM) [23]; and a regularized mixture model (RMM) [20]. The main finding of this paper was that, in general, RM3 is the best and most stable method among the others.…”
Section: Introductionmentioning
confidence: 93%
“…The main finding of this paper was that, in general, RM3 is the best and most stable method among the others. RM3 and RM4 [1] are extensions of the originally formulated RM1 and RM2 approximations, respectively. These extensions linearly interpolate the original query with the terms selected for expansion using RM1 or RM2.…”
Section: Introductionmentioning
confidence: 99%
“…F-score Blott et al (2004) / Gamon (2006 0.622 Tomiyama et al (2004) 0.619 Abdul-Jaleel et al (2004) 0.618 Schiffman and McKeown (2004) 0 …”
Section: Systemsmentioning
confidence: 99%