Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval 2022
DOI: 10.1145/3477495.3531833
|View full text |Cite
|
Sign up to set email alerts
|

An Efficiency Study for SPLADE Models

Abstract: Latency and efficiency issues are often overlooked when evaluating IR models based on Pretrained Language Models (PLMs) in reason of multiple hardware and software testing scenarios. Nevertheless, efficiency is an important part of such systems and should not be overlooked.In this paper, we focus on improving the efficiency of the SPLADE model since it has achieved state-of-the-art zero-shot performance and competitive results on TREC collections. SPLADE efficiency can be controlled via a regularization factor… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
15
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 36 publications
(16 citation statements)
references
References 45 publications
1
15
0
Order By: Relevance
“…We confirmed this by replacing the shared encoder with two separate ones (distilSplade sep ), which reduced latency from 122.5 ms to 50.2 ms, a 59% decrease. This benefit of separate encoders was also reported in [14], and our results further support its substantial impact.…”
Section: Rq2: How Do Lsr Methods Perform With Recent Advanced Trainin...supporting
confidence: 88%
See 1 more Smart Citation
“…We confirmed this by replacing the shared encoder with two separate ones (distilSplade sep ), which reduced latency from 122.5 ms to 50.2 ms, a 59% decrease. This benefit of separate encoders was also reported in [14], and our results further support its substantial impact.…”
Section: Rq2: How Do Lsr Methods Perform With Recent Advanced Trainin...supporting
confidence: 88%
“…In practice, distilSplade qM LP could be viewed as a more efficient drop-in replacement for the full model. This use of qM LP is complementary to other changes (e.g., using a smaller encoder as in [14]) to improve the efficiency of LSR.…”
Section: Rq3: How Does the Choice Of Encoder Architecture And Regular...mentioning
confidence: 99%
“…Efficient SPLADE model. Table 5 shows the application of 2GTI in a recently published efficient SPLADE model [20] which has made several improvements in retrieval speed. We have used the released checkpoint of this efficient model called BT-SPLADE-L, which has a weaker MRR@10 score, but significantly faster than our trained SPLADE baseline reported in Table 2.…”
Section: Evaluationsmentioning
confidence: 99%
“…We follow the strategy we used on our latest TREC notebooks, in that we strive for making this more streamlined than a normal research paper would be. We will now present a list of the papers that better introduce and detail the models we used here and refer the reader to check them for a better explanation than those we have here, that are mainly dedicated to how to apply it to MIRACL and not to the methods themselves: i) Training non English SPLADE models [11], ii) The SPLADE model [5,10], iii) The Contriever model and its pretraining [8], iv) The RankT5 reranker [16], v) MonoT5 [13], vi) The LCE loss [6], vii) ColBERT [9], and viii) For our ensembling we use Ranx [1] and their min-max normalized sum ensembling.…”
Section: Introductionmentioning
confidence: 99%
“…This was mostly due to desperation when we saw everyone overtaking us on the dev set. These rerankers were then tested on known languages and they did not improve the results 10. https://huggingface.co/google/byt5-xl 11 https://huggingface.co/microsoft/mdeberta-v3-base…”
mentioning
confidence: 99%