Proceedings of the ACM Web Conference 2022 2022
DOI: 10.1145/3485447.3511978
|View full text |Cite
|
Sign up to set email alerts
|

Semi-Siamese Bi-encoder Neural Ranking Model Using Lightweight Fine-Tuning

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
2
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(8 citation statements)
references
References 6 publications
0
8
0
Order By: Relevance
“…Thus, our PEFA is also applicable to both pre-trained and fine-tuned ERMs, even ones initialized from black-box LLMs. Note that PEFA is orthogonal and complement to most existing literature that aims to obtain better pre-trained or fine-tuned ERMs at the learning stage, including recent studies of the parameter-efficient fine-tuning of ERMs [28,37,44]. Finally, for the ease of discussion, we assume embeddings obtained from ERMs are unit-norm (i.e., ℓ 2 normalized), hence the inner product is equivalent to the cosine similarity.…”
Section: Problem Statementmentioning
confidence: 99%
“…Thus, our PEFA is also applicable to both pre-trained and fine-tuned ERMs, even ones initialized from black-box LLMs. Note that PEFA is orthogonal and complement to most existing literature that aims to obtain better pre-trained or fine-tuned ERMs at the learning stage, including recent studies of the parameter-efficient fine-tuning of ERMs [28,37,44]. Finally, for the ease of discussion, we assume embeddings obtained from ERMs are unit-norm (i.e., ℓ 2 normalized), hence the inner product is equivalent to the cosine similarity.…”
Section: Problem Statementmentioning
confidence: 99%
“…in a Siamese fashion. Jung et al [15] use a semi-Siamese setting, where the encoders do share parameters as well, but they are adapted to their specific role (query or document encoding) using light fine-tuning methods. We are not aware of any approaches that employ heterogeneous models, where the two encoders do not share the same model architecture and initial weights.…”
Section: Neural Retrieval and Rankingmentioning
confidence: 99%
“…(1) The encoders may be unable to fully adapt to the characteristics of their respective inputs. For example, queries are usually short and concise whereas documents are longer and more complex [15]. (2) The query encoder has the same number of parameters as the document encoder by design.…”
Section: Heterogeneous Dual-encodersmentioning
confidence: 99%
See 2 more Smart Citations