2022
DOI: 10.48550/arxiv.2201.12438
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Commonsense Knowledge Reasoning and Generation with Pre-trained Language Models: A Survey

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 0 publications
0
5
0
Order By: Relevance
“…4000 sentence pairs Crowd sourcing CosmosQA [63] Question answering 35,600 problems Crowd sourcing CommonsenseQA [130] Question answering 12,247 questions Crowd sourcing CommonsenseQA 2.0 [131] Yes/no questions 14,343 questions Gamification COPA [111] Select a conclusion causally 1000 questions Expert authors connected to a premise CREAK [104] True/False questions 13,000 questions Crowd sourcing CycIC (No paper) Question answering 10,700 questions Synthesized DefeasibleNLI [114] Does new information 250,000 examples Crowd sourcing strenghthen an inference? DiscoSense [11] Understanding of 13,056 examples Adapted discourse markers ETHICS [59] Ethical judgments 130,000+ examples Crowd sourcing FeasibilityQA [57] Question answering 4608 questions Expert construction. GeoMLAMA [150] Question answering 3145 questions Expert construction Multilingual en zh hi fa sw HellaSwag [156] What will 70,000 questions Adapted & happen next?…”
Section: Collections Of Elementary Mathematical Word Problemsmentioning
confidence: 99%
See 1 more Smart Citation
“…4000 sentence pairs Crowd sourcing CosmosQA [63] Question answering 35,600 problems Crowd sourcing CommonsenseQA [130] Question answering 12,247 questions Crowd sourcing CommonsenseQA 2.0 [131] Yes/no questions 14,343 questions Gamification COPA [111] Select a conclusion causally 1000 questions Expert authors connected to a premise CREAK [104] True/False questions 13,000 questions Crowd sourcing CycIC (No paper) Question answering 10,700 questions Synthesized DefeasibleNLI [114] Does new information 250,000 examples Crowd sourcing strenghthen an inference? DiscoSense [11] Understanding of 13,056 examples Adapted discourse markers ETHICS [59] Ethical judgments 130,000+ examples Crowd sourcing FeasibilityQA [57] Question answering 4608 questions Expert construction. GeoMLAMA [150] Question answering 3145 questions Expert construction Multilingual en zh hi fa sw HellaSwag [156] What will 70,000 questions Adapted & happen next?…”
Section: Collections Of Elementary Mathematical Word Problemsmentioning
confidence: 99%
“…When someone wants to test their AI on your benchmark, they train on your published training set, then they upload their trained model onto your machine, and you can run their model on your test set without exposing its contents. This approach was taken, for instance, in the long series of NLP challenges organized by NIST and by the leaderboards for various tasks administered by the Allen AI Institute 11. Since the training set and test set are random samples of the same corpus, they should have essentially identical features.…”
mentioning
confidence: 99%
“…Another line of work proposes the evaluation of LLMs on downstream tasks exploiting well-crafted datasets, each of which is dedicated on different reasoning senses. To this end, various tests have stressed LLM capabilities on arithmetic [59], symbolic [58], commonsense [60], and other types of reasoning. Overall, the findings occurring from the aforementioned endeavors suggest that indeed, LLMs present emergent reasoning capabilities simulating human thinking patterns, though being incapable of tackling complex reasoning challenges.…”
Section: Reasoning In Knowledge Graphs and Large Language Modelsmentioning
confidence: 99%
“…After pretraining a language model (LM), it undoubtedly acquires linguistic knowledge but also factual (facts about entities), relational (relation between concepts/entities), and commonsense knowledge [32,35]. However, they also tend to suffer from some issues such as (1) catastrophic forgetting after fine-tuning on a downstream task despite any type of regularization, (2) the hallucination problem, for example in dialogue systems (pretrained LMs generate factually incorrect statements) [8,34], and (3) the fact that they rely on memorization during pretraining, which makes them struggle with unseen entities [4,21]. To alleviate these issues, a certain number of recent approaches focused on solutions to better implicitly incorporate knowledge in Pretrained Language Models (PLMs).…”
Section: Knowledge Injection Into Language Modelsmentioning
confidence: 99%