2021
DOI: 10.48550/arxiv.2109.03564
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

NSP-BERT: A Prompt-based Zero-Shot Learner Through an Original Pre-training Task--Next Sentence Prediction

Abstract: Using prompts to utilize language models to perform various downstream tasks, also known as prompt-based learning or prompt-learning, has lately gained significant success in comparison to the pre-train and fine-tune paradigm. Nonetheless, virtually all prompt-based methods are tokenlevel, meaning they all utilize GPT's left-to-right language model or BERT's masked language model to perform clozestyle tasks. In this paper, we attempt to accomplish several NLP tasks in the zero-shot scenario using a BERT origin… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
24
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
4

Relationship

0
10

Authors

Journals

citations
Cited by 21 publications
(24 citation statements)
references
References 35 publications
0
24
0
Order By: Relevance
“…Wang et al (2021) transform NLP tasks into textual entailment and provide label-specific descriptions for each class. Sun et al (2021) propose an approach named NSP-BERT which utilizes a BERT original next sentence prediction pre-training task to perform few-shot learning. Additionally, Puri and Catanzaro (2019) show that reformulating NLP tasks as question answering problems to query generative language models is also a feasible approach.…”
Section: Prompt-based Few-shot Learningmentioning
confidence: 99%
“…Wang et al (2021) transform NLP tasks into textual entailment and provide label-specific descriptions for each class. Sun et al (2021) propose an approach named NSP-BERT which utilizes a BERT original next sentence prediction pre-training task to perform few-shot learning. Additionally, Puri and Catanzaro (2019) show that reformulating NLP tasks as question answering problems to query generative language models is also a feasible approach.…”
Section: Prompt-based Few-shot Learningmentioning
confidence: 99%
“…It teaches the model to understand dependencies across sentences [53]. In spite of that, NSP is criticized as a weak task for its comparison of similarity [83]. To overcome this limitation, we introduce a harder snapshot ordering task, which aims to order a set of conformations as a coherent sub-trajectory.…”
Section: Snapshot Ordering Pre-trainingmentioning
confidence: 99%
“…ā€¢ For the natural language inference task, we exploit NSP-based prompt training [100]. Different labels are regarded as prompts to concatenate the two sentences, and the model is trained to select the label that makes the concatenated sentence the most coherent.…”
Section: Settingsmentioning
confidence: 99%