Proceedings of Deep Learning Inside Out (DeeLIO 2022): The 3rd Workshop on Knowledge Extraction and Integration for Deep Learni 2022
DOI: 10.18653/v1/2022.deelio-1.10
|View full text |Cite
|
Sign up to set email alerts
|

What Makes Good In-Context Examples for GPT-3?

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

6
204
1

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 255 publications
(211 citation statements)
references
References 24 publications
6
204
1
Order By: Relevance
“…More recently, prompting or masked infilling has been used to probe models for knowledge (Petroni et al, 2019) or perform a variety of NLP tasks (Radford et al, 2019;Brown et al, 2020). There has also been work on eliciting prompting behavior in smaller models (Schick and Schütze, 2020; 11 https://huggingface.co/bigscience/ tr11-176B-ml-logs/tensorboard Gao et al, 2021b;Li and Liang, 2021;Lester et al, 2021;Scao and Rush, 2021), improving the flexibility of prompting (Shin et al, 2020), and understanding why and how prompting works (Liu et al, 2021;Min et al, 2022).…”
Section: Limitationsmentioning
confidence: 99%
“…More recently, prompting or masked infilling has been used to probe models for knowledge (Petroni et al, 2019) or perform a variety of NLP tasks (Radford et al, 2019;Brown et al, 2020). There has also been work on eliciting prompting behavior in smaller models (Schick and Schütze, 2020; 11 https://huggingface.co/bigscience/ tr11-176B-ml-logs/tensorboard Gao et al, 2021b;Li and Liang, 2021;Lester et al, 2021;Scao and Rush, 2021), improving the flexibility of prompting (Shin et al, 2020), and understanding why and how prompting works (Liu et al, 2021;Min et al, 2022).…”
Section: Limitationsmentioning
confidence: 99%
“…To get the demonstrations, GPT-3 uses a random strategy to drawn from D train . Gao et al (2021) and Liu et al (2021a) use a BERT (Devlin et al, 2019)/RoBERTa (Liu et al, 2019) based retriever to sample examples with similarity. Their experiment 1 D train represents the training dataset for classification problems in their paper.…”
Section: Demonstration With Similaritymentioning
confidence: 99%
“…This method is first used by GPT series (Radford et al, 2019;Brown et al, 2020). While in GPT models the demonstrations are selected randomly, researchers found that the selection with similarity would significantly improve the final performance (Gao et al, 2021;Liu et al, 2021a). Liu et al (2021a) and Kumar and Talukdar (2021) also discovered that the order of prompts provided to the model has a great influence on the performance of the model.…”
Section: Generality Testmentioning
confidence: 99%
See 1 more Smart Citation
“…Neural vs. Discrete Demonstration Compared with prior discrete demonstrations described in [12,33,47,26], retrieving weighted neural demonstrations from the knowledge-store to augment prompt learning has advantages in the following three major aspects: (1) neural demonstrations could be more tolerant of the model's maximum input length than discrete demonstrations, while the discrete demonstration is usually not suitable for multi-class classification tasks due to the limitation of input length, such as relation extraction, etc. (2) the model needs to deal with large retrieval tokens for discrete demonstration, making it time-consuming and computationally intensive to perform cross-attention operations due to the quadratic attention complexity.…”
Section: Retrieval Of Neural Demonstrationmentioning
confidence: 99%