2020
DOI: 10.48550/arxiv.2011.00777
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

COSMO: Conditional SEQ2SEQ-based Mixture Model for Zero-Shot Commonsense Question Answering

Abstract: Commonsense reasoning refers to the ability of evaluating a social situation and acting accordingly. Identification of the implicit causes and effects of a social context is the driving capability which can enable machines to perform commonsense reasoning. The dynamic world of social interactions requires context-dependent on-demand systems to infer such underlying information. However, current approaches in this realm lack the ability to perform commonsense reasoning upon facing an unseen situation, mostly du… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 13 publications
0
2
0
Order By: Relevance
“…Specifically, by designing and constructing an appropriate input sequence of words (called a prompt), one is able to induce the model to produce the desired output sequence (i.e., a paraphrased utterance in this context)without changing its parameters through fine-tuning at all. In particular, in the low-data regime, empirical analysis shows that, either for manually picking hand-crafted prompts [21] or automatically building auto-generated prompts [8], [16] taking prompts for tuning models is surprisingly effective for the knowledge stimulation and model adaptation of pre-trained language model. However, none of these methods reports the use of a few-shot pre-trained language model to directly generate few-shot and zero-shot paraphrased utterances.…”
Section: Paraphrase Generationmentioning
confidence: 99%
See 1 more Smart Citation
“…Specifically, by designing and constructing an appropriate input sequence of words (called a prompt), one is able to induce the model to produce the desired output sequence (i.e., a paraphrased utterance in this context)without changing its parameters through fine-tuning at all. In particular, in the low-data regime, empirical analysis shows that, either for manually picking hand-crafted prompts [21] or automatically building auto-generated prompts [8], [16] taking prompts for tuning models is surprisingly effective for the knowledge stimulation and model adaptation of pre-trained language model. However, none of these methods reports the use of a few-shot pre-trained language model to directly generate few-shot and zero-shot paraphrased utterances.…”
Section: Paraphrase Generationmentioning
confidence: 99%
“…3) Paraphrasing using fine-tuned language model generation: Previous researches have proposed several pretrained autoregressive language models, such as GPT-2 [28] and ProphetNet [26]. Other studies such as [21] use these language models to solve various downstream tasks and achieve SOTA performance. In this work, following [41], we fine-tune an autoregressive language model BART [15] on the dataset from [14], which is a subset of the PARANMT corpus [38], to generate syntactically and lexically diversified paraphrases.…”
Section: Paraphrase Generationmentioning
confidence: 99%