Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2022
DOI: 10.18653/v1/2022.naacl-main.272
|View full text |Cite
|
Sign up to set email alerts
|

ColBERTv2: Effective and Efficient Retrieval via Lightweight Late Interaction

Abstract: Neural information retrieval (IR) has greatly advanced search and other knowledgeintensive language tasks. While many neural IR methods encode queries and documents into single-vector representations, late interaction models produce multi-vector representations at the granularity of each token and decompose relevance modeling into scalable token-level computations. This decomposition has been shown to make late interaction more effective, but it inflates the space footprint of these models by an order of magni… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
61
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 102 publications
(61 citation statements)
references
References 51 publications
0
61
0
Order By: Relevance
“…Dense passage retrieval Khattab and Zaharia, 2020;Luan et al, 2021;Santhanam et al, 2021) has gained a lot of attention lately with applications extending beyond retrieval tasks into areas including open-domain question answering, language model pre-training, fact checking, dialogue generation (e.g., RAG , REALM (Guu et al, 2020), MultiDPR , KILT , Con-vDR , RocketQA (Qu et al, 2021)). In dense passage retrieval, the query q and each passage p are separately encoded into dense vectors, and relevance is modeled via similarity functions such as dot-product.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…Dense passage retrieval Khattab and Zaharia, 2020;Luan et al, 2021;Santhanam et al, 2021) has gained a lot of attention lately with applications extending beyond retrieval tasks into areas including open-domain question answering, language model pre-training, fact checking, dialogue generation (e.g., RAG , REALM (Guu et al, 2020), MultiDPR , KILT , Con-vDR , RocketQA (Qu et al, 2021)). In dense passage retrieval, the query q and each passage p are separately encoded into dense vectors, and relevance is modeled via similarity functions such as dot-product.…”
Section: Related Workmentioning
confidence: 99%
“…Dense passage retrieval methods Khattab and Zaharia, 2020;Santhanam et al, 2021;Luan et al, 2021;Humeau et al, 2020;MacAvaney et al, 2020; have gained a lot of attention lately and achieved state of the art results on various retrieval and ranking datasets. Dense retrievers are efficient compared to other neural methods such as transformer-based cross-encoder models: passages are encoded and indexed offline, at inference time only the query needs to be encoded once; also they leverage ANN (approximate nearest neighbor) algorithms to efficiently search for relevant dense vectors.…”
Section: Task and Baselinesmentioning
confidence: 99%
See 2 more Smart Citations
“…Dense passage retrieval Luan et al, 2021;Santhanam et al, 2021) has gained a lot of attention lately with applications extending beyond retrieval tasks into areas including open-domain question answering, language model pre-training, fact checking, dialogue generation (e.g., RAG , REALM (Guu et al, 2020), MultiDPR , KILT , Con-vDR , RocketQA (Qu et al, 2021)). In dense passage retrieval, the query q and each passage p are separately encoded into dense vectors, and relevance is modeled via similarity functions such as dot-product.…”
Section: Related Workmentioning
confidence: 99%