2021
DOI: 10.48550/arxiv.2110.03141
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Efficient Sharpness-aware Minimization for Improved Training of Neural Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
25
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 12 publications
(25 citation statements)
references
References 13 publications
0
25
0
Order By: Relevance
“…Recently, sharpness-aware minimization (SAM) [16] seeks to find parameters that lie in a region with both low loss value and loss sharpness and shows promising performance across various architectures and benchmark datasets. Moreover, several methods have been proposed to improve the performance [31] or efficiency [13] of SAM. Specifically, ASAM [31] introduces a concept of adaptive sharpness to mitigate the effect of parameter re-scaling while ESAM [13] reduces the computational overhead without performance drop.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Recently, sharpness-aware minimization (SAM) [16] seeks to find parameters that lie in a region with both low loss value and loss sharpness and shows promising performance across various architectures and benchmark datasets. Moreover, several methods have been proposed to improve the performance [31] or efficiency [13] of SAM. Specifically, ASAM [31] introduces a concept of adaptive sharpness to mitigate the effect of parameter re-scaling while ESAM [13] reduces the computational overhead without performance drop.…”
Section: Related Workmentioning
confidence: 99%
“…Moreover, several methods have been proposed to improve the performance [31] or efficiency [13] of SAM. Specifically, ASAM [31] introduces a concept of adaptive sharpness to mitigate the effect of parameter re-scaling while ESAM [13] reduces the computational overhead without performance drop. Compared with these existing methods, our proposed SAQ focuses on improving the generalization performance of the quantized models.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…SAM explicitly penalizes a sharpness measure to obtain flat minima, which has achieved state-of-the-art results in several learning tasks [2,37]. ESAM [6], GSAM [37], and SAF. GSAM is the stateof-the-art among SAM's follow-up works.…”
Section: Introductionmentioning
confidence: 99%
“…Liu et al [21] and Du et al [6] recently addressed the computation issue of SAM and proposed LookSAM [21] and Efficient SAM (ESAM) [6], respectively. LookSAM only minimizes the sharpness measure once in the first of every five iterations.…”
Section: Introductionmentioning
confidence: 99%