2022
DOI: 10.48550/arxiv.2212.08410
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Teaching Small Language Models to Reason

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
24
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 14 publications
(24 citation statements)
references
References 0 publications
0
24
0
Order By: Relevance
“…Our work exploits generating two types of teaching data from LLMs and mixing teaching data. We discover that this simple method highly improves student performance in complex multi-modality tasks, which has not yet been recognized in previous studies on fine-tuning with CoT reasoning (Hsieh et al 2023;Ho, Schmid, and Yun 2022;Huang et al 2022;Magister et al 2022;Fu et al 2023;Hu et al 2023).…”
Section: Related Workmentioning
confidence: 54%
See 1 more Smart Citation
“…Our work exploits generating two types of teaching data from LLMs and mixing teaching data. We discover that this simple method highly improves student performance in complex multi-modality tasks, which has not yet been recognized in previous studies on fine-tuning with CoT reasoning (Hsieh et al 2023;Ho, Schmid, and Yun 2022;Huang et al 2022;Magister et al 2022;Fu et al 2023;Hu et al 2023).…”
Section: Related Workmentioning
confidence: 54%
“…In recent studies, CoT reasoning is elicited in small models using fine-tuned language models. Magister et al (2022) benefit smaller models through CoT distillation. Huang et al (2022) show that LLMs can enhance reasoning using self-generated solutions from unlabeled data.…”
Section: Related Workmentioning
confidence: 99%
“…Some recent works also tackle the problem by distilling LLMs into smaller LMs for domain-specific tasks (Magister et al 2022;Hsieh et al 2023;Wang et al 2022a. Wang et al (2022a) propose PINTO, which uses LLMs' rationales as context of a small LM to improve its performance on reasoning tasks.…”
Section: Distilling Llmsmentioning
confidence: 99%
“…further propose SCOTT, a faithful knowledge distillation method that elicits self-consistent rationales from LLMs to fine-tune a small while more faithful LM for open-domain questionanswering tasks. Both Magister et al (2022) and Hsieh et al (2023) demonstrate that LLMs can be distilled into smaller but more effective LMs by fine-tuning with both answers and rationales on commonsense reasoning and arithmetic tasks. Different from these works, our work focus on the domain of toxic content detection.…”
Section: Distilling Llmsmentioning
confidence: 99%
“…This opens up new avenues of research, referred to as prompt engineering and meta-learning. Recent studies [11], [13], [18]- [20] highlight the significance and relevance of exploring these areas.…”
Section: Related Workmentioning
confidence: 99%