2022
DOI: 10.48550/arxiv.2205.11194
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

UnifieR: A Unified Retriever for Large-Scale Retrieval

Abstract: Large-scale retrieval is to recall relevant documents from a huge collection given a query. It relies on representation learning to embed documents and queries into a common semantic encoding space. According to the encoding space, recent retrieval methods based on pre-trained language models (PLM) can be coarsely categorized into either dense-vector or lexicon-based paradigms. These two paradigms unveil the PLMs' representation capability in different granularities, i.e., global sequence-level compression and… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 25 publications
0
2
0
Order By: Relevance
“…[187] proposed a method called Representational Slicing which densifies sparse representations for passage retrieval. [188]) proposed a method called UnifieR which aims to improve the efficiency and effectiveness of large-scale retrieval by unifying different retrieval strategies.…”
Section: Multi-vector Representationmentioning
confidence: 99%
“…[187] proposed a method called Representational Slicing which densifies sparse representations for passage retrieval. [188]) proposed a method called UnifieR which aims to improve the efficiency and effectiveness of large-scale retrieval by unifying different retrieval strategies.…”
Section: Multi-vector Representationmentioning
confidence: 99%
“…Compared with traditional lexical-based methods, the deep semantic retrieval model has the advantages of high accuracy, low mismatch, and strong generalization. The classical deep semantic retrieval methods could be split into two categories, sparse-based retrieval (Bai et al, 2020;Shen et al, 2022;Formal et al, 2021;Gao et al, 2021), and dense-based retrieval Khattab and Zaharia, 2020;Zhan et al, 2021;Qiu et al, 2022;. Although there are differences in representation, they all learn the deep model through an end-to-end paradigm using contrastive learning methods.…”
Section: Introductionmentioning
confidence: 99%