2020
DOI: 10.48550/arxiv.2007.12223
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The Lottery Ticket Hypothesis for Pre-trained BERT Networks

Abstract: In natural language processing (NLP), enormous pre-trained models like BERT have become the standard starting point for training on a range of downstream tasks, and similar trends are emerging in other areas of deep learning. In parallel, work on the lottery ticket hypothesis has shown that models for NLP and computer vision contain smaller matching subnetworks capable of training in isolation to full accuracy and transferring to other tasks. In this work, we combine these observations to assess whether such t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

2
50
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 44 publications
(52 citation statements)
references
References 21 publications
2
50
0
Order By: Relevance
“…On the other hand, in structured pruning, the best subnetworks of BERT's heads do not quite reach the full model performance. (Chen et al, 2020a) shows for a range of downstream tasks, matching subnetworks at 40% to 90% sparsity exist and they are found at pretrained phase (initialization). This is dissimilar to the prior NLP research where subnetworks emerge only after some amount of training.…”
Section: The Lottery Ticket Hypothesismentioning
confidence: 99%
See 3 more Smart Citations
“…On the other hand, in structured pruning, the best subnetworks of BERT's heads do not quite reach the full model performance. (Chen et al, 2020a) shows for a range of downstream tasks, matching subnetworks at 40% to 90% sparsity exist and they are found at pretrained phase (initialization). This is dissimilar to the prior NLP research where subnetworks emerge only after some amount of training.…”
Section: The Lottery Ticket Hypothesismentioning
confidence: 99%
“…In view of that, follow-up works reveal that sparsity patterns might emerge at the initialization , the early stage of training (You et al, 2019) and (Chen et al, 2020b), or in dynamic forms throughout training (Evci et al, 2020) by updating model parameters and architecture typologies simultaneously. Some of the recent findings are that the lottery ticket hypothesis holds for BERT models, i.e., largest weights of the original network do form subnetworks that can be retrained alone to reach the performance close to that of the full model (Prasanna et al, 2020;Chen et al, 2020a).…”
Section: The Lottery Ticket Hypothesismentioning
confidence: 99%
See 2 more Smart Citations
“…To address this issue, many efforts have been devoted to compressing the cumbersome transformer architectures into a lightweight counterpart, including knowledge distillation (Jiao et al, 2019;Sanh et al, 2019;Sun et al, 2019;, pruning (Michel et al, 2019;Chen et al, 2020;, and quantization (Zafrir et al, 2019;Bai et al, 2020;Shen et al, 2020). Among all these compression techniques, quantization is a popular solution as it still preserves the original network architecture.…”
Section: Introductionmentioning
confidence: 99%