Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-long.393
|View full text |Cite
|
Sign up to set email alerts
|

ConSERT: A Contrastive Framework for Self-Supervised Sentence Representation Transfer

Abstract: Learning high-quality sentence representations benefits a wide range of natural language processing tasks. Though BERT-based pretrained language models achieve high performance on many downstream tasks, the native derived sentence representations are proved to be collapsed and thus produce a poor performance on the semantic textual similarity (STS) tasks. In this paper, we present ConSERT, a Contrastive Framework for Self-Supervised SEntence Representation Transfer, that adopts contrastive learning to fine-tun… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
130
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 288 publications
(181 citation statements)
references
References 23 publications
0
130
0
Order By: Relevance
“…Pagliardini et al (2018) show that simply augmenting the idea of word2vec (Mikolov et al, 2013) with n-gram embeddings leads to strong results. Several recent (and concurrent) approaches adopt contrastive objectives (Zhang et al, 2020;Giorgi et al, 2021;Meng et al, 2021;Carlsson et al, 2021;Kim et al, 2021;Yan et al, 2021) by taking different views-from data augmentation or different copies of models-of the same sentence or document. Compared to these work, SimCSE uses the simplest idea by taking different outputs of the same sentence from standard dropout, and performs the best on STS tasks.…”
Section: Related Workmentioning
confidence: 99%
“…Pagliardini et al (2018) show that simply augmenting the idea of word2vec (Mikolov et al, 2013) with n-gram embeddings leads to strong results. Several recent (and concurrent) approaches adopt contrastive objectives (Zhang et al, 2020;Giorgi et al, 2021;Meng et al, 2021;Carlsson et al, 2021;Kim et al, 2021;Yan et al, 2021) by taking different views-from data augmentation or different copies of models-of the same sentence or document. Compared to these work, SimCSE uses the simplest idea by taking different outputs of the same sentence from standard dropout, and performs the best on STS tasks.…”
Section: Related Workmentioning
confidence: 99%
“…CERT (Fang and Xie, 2020) mainly using back-and-forth translation, and CLINE proposed synonym substitution as positive samples and antonym substitution as negative samples, and then minimize the triplet loss between positive, negative cases as well as the original text. ConSERT (Yan et al, 2021) uses adversarial attack, token shuffling, cutoff, and dropout as data augmentation. CLAE (Ho and Nvasconcelos, 2020) also introduces Fast Gradient Sign Method, an adversarial attack method, as text data augmentation.…”
Section: Related Workmentioning
confidence: 99%
“…Specifically, Reimers and Gurevych (2019) mainly use the classification objective for an NLI dataset, and adopt contrastive learning to utilize self-supervision from a large corpus. Yan et al (2021); Gao et al (2021) incorporate a parallel corpus such as NLI datasets into their contrastive learning framework.…”
Section: Semantic Textual Similaritymentioning
confidence: 99%