Proceedings of the 31st ACM International Conference on Information &Amp; Knowledge Management 2022
DOI: 10.1145/3511808.3557067
|View full text |Cite
|
Sign up to set email alerts
|

e-CLIP: Large-Scale Vision-Language Representation Learning in E-commerce

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(1 citation statement)
references
References 19 publications
0
1
0
Order By: Relevance
“…Recently, with the emergence of datasets that support various modalities, studies using various modality information have emerged. Shin et al [ 25 ] proposed e-CLIP, which can be deployed on multiple e-commerce downstream tasks, based on an approach [ 26 ] that utilizes both visual and language information. Dong et al [ 13 ] proposed the Self-harmonized Contrastive Learning (SCALE) framework, which unifies the several modalities into a unified model through an adaptive mechanism for fusing features.…”
Section: Related Workmentioning
confidence: 99%
“…Recently, with the emergence of datasets that support various modalities, studies using various modality information have emerged. Shin et al [ 25 ] proposed e-CLIP, which can be deployed on multiple e-commerce downstream tasks, based on an approach [ 26 ] that utilizes both visual and language information. Dong et al [ 13 ] proposed the Self-harmonized Contrastive Learning (SCALE) framework, which unifies the several modalities into a unified model through an adaptive mechanism for fusing features.…”
Section: Related Workmentioning
confidence: 99%