2018 IEEE International Symposium on Information Theory (ISIT) 2018
DOI: 10.1109/isit.2018.8437571
|View full text |Cite
|
Sign up to set email alerts
|

Generalization Error Bounds for Noisy, Iterative Algorithms

Abstract: In statistical learning theory, generalization error is used to quantify the degree to which a supervised machine learning algorithm may overfit to training data. Recent work [Xu and Raginsky (2017)] has established a bound on the generalization error of empirical risk minimization based on the mutual information I(S; W ) between the algorithm input S and the algorithm output W , when the loss function is sub-Gaussian. We leverage these results to derive generalization error bounds for a broad class of iterati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

4
60
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 60 publications
(64 citation statements)
references
References 8 publications
4
60
0
Order By: Relevance
“…The bound in ( 45 ) has the same form as the generalization gap derived in [ 27 ] for conventional learning. From ( 45 ), the generalization gap can be reduced by increasing the variance of the injected Gaussian noise.…”
Section: Applicationsmentioning
confidence: 91%
See 3 more Smart Citations
“…The bound in ( 45 ) has the same form as the generalization gap derived in [ 27 ] for conventional learning. From ( 45 ), the generalization gap can be reduced by increasing the variance of the injected Gaussian noise.…”
Section: Applicationsmentioning
confidence: 91%
“…The sampling strategy in ( 44 ) together with ( A19 ) then implies the following relation Using to denote the set of all updates, we have the following relations where, the inequality in (a) follows from data processing inequality on Markov chain ; (b) follows from the Markov chain ; and the equality in (c) follows from and ( A20 ). Finally, the computation of bound in ( A24 ) follows similar to Lemma 5 in [ 27 ].…”
Section: Definitionmentioning
confidence: 99%
See 2 more Smart Citations
“…Hardt et al (2015) first established random uniform stability for randomized iterative SGD algorithms for convex and non-convex settings in pointwise learning. The results were further improved in the work (Kuzborskij and Lampert, 2017;Pensia et al, 2018) by exploring the structures of the loss function and the data.…”
Section: Introductionmentioning
confidence: 99%