Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers) 2022
DOI: 10.18653/v1/2022.acl-short.8
|View full text |Cite
|
Sign up to set email alerts
|

P-Tuning: Prompt Tuning Can Be Comparable to Fine-tuning Across Scales and Tasks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
210
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 399 publications
(210 citation statements)
references
References 0 publications
0
210
0
Order By: Relevance
“…In this section, we show that DEEPSTRUCT successfully transfers to the structure prediction tasks considered and obtain state-of-the-art results on 21 of 28 datasets we evaluate. All results are obtained via structure pretraining a pretrained 10B parameter LM, GLM (Du et al, 2021). The details of the experimental setup, datasets, and comparison methods are described in Appendix A.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…In this section, we show that DEEPSTRUCT successfully transfers to the structure prediction tasks considered and obtain state-of-the-art results on 21 of 28 datasets we evaluate. All results are obtained via structure pretraining a pretrained 10B parameter LM, GLM (Du et al, 2021). The details of the experimental setup, datasets, and comparison methods are described in Appendix A.…”
Section: Methodsmentioning
confidence: 99%
“…We adopt the pretrained LMs from the (Du et al, 2021), whose energy cost and carbon footprint during pretraining were 80.6 MWh and 4.6 tCO2e, respectively. Additionally, the structure pretraining takes less than 5% gradient-steps of the number of pretraining steps of LMs, and thus the estimated auxiliary cost for energy is comparatively smaller.…”
Section: Environmental Considerationsmentioning
confidence: 99%
“…Since the summer of 2021 there has been a steady influx of research papers concerning prompt learning for common benchmarking open-NLP datasets such as Stanford Sentiment Treebank v2 (SST2), and the General Language Understanding Evaluation (GLUE) Liu et al [2021a], Brown et al [2020b], Sanh et al [2022], Lester et al [2021], Liu et al [2021b], Li and Liang [2021]. The datasets and tasks are standard in the field of NLP, and revolve around natural language understanding (NLU) tasks.…”
Section: Related Workmentioning
confidence: 99%
“…The common finding is that prompt learning can reach the performance of traditional fine-tuning, and often outperform in few-shot settings. Although the ability of prompt learning to match performance of traditional fine-tuning seems to scale with PLM size Liu et al [2021b]. One notable paper has investigated the use of GPT-3 for biomedical text datasets in a few-shot setting, finding a decrease in performance when compared to similar tasks in the standard NLU datasets Moradi et al [2021].…”
Section: Related Workmentioning
confidence: 99%
“…Lester et al (2021) propose a further simplified approach called prompt tuning, which only tunes the additional tunable tokens prepended to the input text. P-tuning v2 (Liu et al, 2021b) adapted the idea of prompt tuning by adding prompts in different layers as pre-fix tokens rather than only the input embedding. Its performance can be comparable to full-model tuning across both scales and tasks.…”
Section: Related Workmentioning
confidence: 99%