Findings of the Association for Computational Linguistics: NAACL 2022 2022
DOI: 10.18653/v1/2022.findings-naacl.25
|View full text |Cite
|
Sign up to set email alerts
|

PCEE-BERT: Accelerating BERT Inference via Patient and Confident Early Exiting

Abstract: BERT and other pre-trained language models (PLMs) are ubiquitous in modern NLP. Even though PLMs are the state-of-the-art (SOTA) models for almost every NLP task (Qiu et al., 2020), the significant latency during inference prohibits wider industrial usage. In this work, we propose Patient and Confident Early Exiting BERT (PCEE-BERT), an off-the-shelf sample-dependent early exiting method that can work with different PLMs and can also work along with popular model compression methods. With a multi-exit BERT as … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(10 citation statements)
references
References 13 publications
0
10
0
Order By: Relevance
“…In PABEE (Zhou et al 2020), the instance exits when k consecutive internal classifiers make the same prediction. PCEE-BERT (Zhang et al 2022) combined both ensemble-based and confidence-based exiting criteria. The instance exits if the confidence scores are greater than a predefined threshold for several consecutive exits.…”
Section: Related Work Early Exitingmentioning
confidence: 99%
See 3 more Smart Citations
“…In PABEE (Zhou et al 2020), the instance exits when k consecutive internal classifiers make the same prediction. PCEE-BERT (Zhang et al 2022) combined both ensemble-based and confidence-based exiting criteria. The instance exits if the confidence scores are greater than a predefined threshold for several consecutive exits.…”
Section: Related Work Early Exitingmentioning
confidence: 99%
“…The Thirty-Eighth AAAI Conference on Artificial Intelligence the sum of cross-entropy losses. SkipBERT (Wang et al 2022), PABEE (Zhou et al 2020), Past-Future (Liao et al 2021), PCEE-BERT (Zhang et al 2022), and LeeBERT (Zhu 2021)) used weighted sum of cross entropy losses. Dee-BERT (Xin et al 2020), Right-Tool (Schwartz et al 2020), BERxiT (Xin et al 2021), and CAT (Schuster et al 2021) used sum of cross entropy losses.…”
Section: Related Work Early Exitingmentioning
confidence: 99%
See 2 more Smart Citations
“…PABEE has proposed a patiencebased exit strategy that halts the forward-pass at an intermediate layer only when the pre-defined number of subsequent layers yield the same predictions. Similarly, DeeBERT and FastBERT have employed the predictive entropy to replace the patience, and PCEE-BERT (Zhang et al, 2022) has combined both patience and confidence for the exit criteria.…”
Section: Depth-wise Reduction On Transformermentioning
confidence: 99%