2023
DOI: 10.1007/978-3-031-33380-4_6
|View full text |Cite
|
Sign up to set email alerts
|

Web-Scale Semantic Product Search with Large Language Models

Abstract: Dense embedding-based semantic matching is widely used in e-commerce product search to address the shortcomings of lexical matching such as sensitivity to spelling variants. The recent advances in BERT-like language model encoders, have however, not found their way to realtime search due to the strict inference latency requirement imposed on e-commerce websites. While bi-encoder BERT architectures enable fast approximate nearest neighbor search, training them effectively on query-product data remains a challen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(2 citation statements)
references
References 14 publications
0
2
0
Order By: Relevance
“…In Table 5, we applied PEFA to pre-trained ERMs (e.g., MPNet base [50], Sentence-T5 base [41], GTR base [42] and E5 base [54]) and the finetuned ERMs (FT-ERM [40]). For privacy of the proprietary product search datasets, we only report the absolute gain of Recall metrics compared to the MPNet base baseline.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…In Table 5, we applied PEFA to pre-trained ERMs (e.g., MPNet base [50], Sentence-T5 base [41], GTR base [42] and E5 base [54]) and the finetuned ERMs (FT-ERM [40]). For privacy of the proprietary product search datasets, we only report the absolute gain of Recall metrics compared to the MPNet base baseline.…”
Section: Resultsmentioning
confidence: 99%
“…ProdSearch-5M ProdSearch-15M ProdSearch-30M Recall@100 Recall@1000 Recall@100 Recall@1000 Recall@100 Recall@1000 MPNet base [50] 0.00 0.00 0.00 0.00 0.00 0.00 +PEFA-XS (ours) [50], Sentence-T5 base [41], GTR base [42] and E5 base [54]) and the finetuned ERMs (FT-ERM [40]) on three proprietary product search datasets: ProdSearch-5M, ProdSearch-15M and ProdSearch-30M.…”
Section: Methodsmentioning
confidence: 99%