2022
DOI: 10.48550/arxiv.2212.09865
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Z-ICL: Zero-Shot In-Context Learning with Pseudo-Demonstrations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
17
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(17 citation statements)
references
References 0 publications
0
17
0
Order By: Relevance
“…As the pseudo-inputs created by SELF-ICL is based on only one reference, i.e., the given test input, the generated pseudo-inputs are likely to be of great semantic similarity with that test input, and fail to capture the correct input space distribution. In such since it has been shown that models tend to copy the labels paired with inputs that are very similar to the test input, known as the copying effect (Lyu et al, 2022). With no guarantee for the correctness of SELF-ICL's pseudo-labels, the copying effect would potentially hurt the ICL performance.…”
Section: The Entanglement Of Input Space and Input-label Mappingmentioning
confidence: 99%
See 1 more Smart Citation
“…As the pseudo-inputs created by SELF-ICL is based on only one reference, i.e., the given test input, the generated pseudo-inputs are likely to be of great semantic similarity with that test input, and fail to capture the correct input space distribution. In such since it has been shown that models tend to copy the labels paired with inputs that are very similar to the test input, known as the copying effect (Lyu et al, 2022). With no guarantee for the correctness of SELF-ICL's pseudo-labels, the copying effect would potentially hurt the ICL performance.…”
Section: The Entanglement Of Input Space and Input-label Mappingmentioning
confidence: 99%
“…Recently, a series of studies has been proposed to shed lights on the inner working of ICL (Xie et al, 2021;Reynolds and McDonell, 2021;Min et al, 2022b). Their evidences suggest that instead of contributing explicit signals for learning new tasks, demonstrations mainly expose large LMs' intrinsic functionalities and guide models towards target domains (Razeghi et al, 2022;Lyu et al, 2022). Similar clues are also partly observed in chain-of-thought (CoT) (Madaan and Yazdanbakhsh, 2022) and instruction-augmented ICL (Webson and Pavlick, 2022).…”
Section: Introductionmentioning
confidence: 99%
“…The performance of ICIL is comparable to adaptive in-context learning methods. We compare ICIL, which samples a fixed demonstration set for all evaluation tasks, with adaptive zero-shot in-context learning (Lyu et al, 2022), where the retrieved demonstrations vary based on the similarity of the target task or instance. Similar to ICIL, for adaptive zero-shot in-context learning, we retrieve the demonstrations which consist of instruction, input, and output instances from the training tasks of SUPERNI benchmark 8 .…”
Section: Additional Experimentsmentioning
confidence: 99%
“…However, Min et al (2022b) show that assigning random labels for demonstrations does not hurt the in-context learning performance much. Motivated by this finding, Lyu et al (2022) propose a zero-shot in-context learning method, retrieving relevant sentences from an external corpus and assigning random labels to construct demonstrations for classification target tasks. Different from Lyu et al (2022), ICIL utilizes instructions to facilitate task adaptation, uses a fixed set of demonstrations to evaluate all tasks, and is applicable to generation target tasks as well.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation