2023
DOI: 10.48550/arxiv.2303.05161
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Inversion dynamics of class manifolds in deep learning reveals tradeoffs underlying generalisation

Abstract: To achieve near-zero training error in a classification problem, the layers of a deep network have to disentangle the manifolds of data points with different labels, to facilitate the discrimination. However, excessive class separation can bring to overfitting since good generalisation requires learning invariant features, which involve some level of entanglement. We report on numerical experiments showing how the optimisation dynamics finds representations that balance these opposing tendencies with a nonmono… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 31 publications
(35 reference statements)
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?