Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2016
DOI: 10.18653/v1/p16-1035
|View full text |Cite
|
Sign up to set email alerts
|

Query Expansion with Locally-Trained Word Embeddings

Abstract: Continuous space word embeddings have received a great deal of attention in the natural language processing and machine learning communities for their ability to model term similarity and other relationships. We study the use of term relatedness in the context of query expansion for ad hoc information retrieval. We demonstrate that word embeddings such as word2vec and GloVe, when trained globally, underperform corpus and query specific embeddings for retrieval tasks. These results suggest that other tasks bene… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

7
179
0
1

Year Published

2017
2017
2020
2020

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 195 publications
(187 citation statements)
references
References 45 publications
7
179
0
1
Order By: Relevance
“…We also compare the results of the proposed approach with a state-of-the-art query expansion approach based on locally trained embeddings [12]. This approach can possibly deal with the problem of multiple degrees of similarity by training the word embeddings on only the top-1000 documents.…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…We also compare the results of the proposed approach with a state-of-the-art query expansion approach based on locally trained embeddings [12]. This approach can possibly deal with the problem of multiple degrees of similarity by training the word embeddings on only the top-1000 documents.…”
Section: Resultsmentioning
confidence: 99%
“…However, the on-line computational overhead could be an issue in practice since the word embeddings are trained on a per-query basis. As only nDCG@10 is used in [12], Table 13 compares the best nDCG@10 reported in [12] with our approach on each of the three publicly available TREC collections. From the comparison results we can see that our method consistently has better results over [12] on all three TREC test collections.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Word embeddings have shown its great practicable usability in plenty of natural language processing tasks, such as information retrieval (Diaz et al, 2016;Zuccon et al, 2015), machine translation (Shi et al, 2016;Zhang et al, 2014), sentiment analysis Xu et al, 2015; and so on. Bilingual word embeddings have been induced for cross-lingual NLP tasks (Vulić and Moens, 2015;Guo et al, 2014;Zou et al, 2013;.…”
Section: Related Workmentioning
confidence: 99%
“…In this session we will focus on semantic matching settings where a supervised signal is available. The signal can be explicit, such as a label for learning task-speci c latent representations [25-27, 36, 47, 48], or relevance labels and, more implicitly, clicks for neural IR methods [15,19,31,41,42]. …”
mentioning
confidence: 99%