2022
DOI: 10.48550/arxiv.2209.04836
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Git Re-Basin: Merging Models modulo Permutation Symmetries

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
14
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 8 publications
(15 citation statements)
references
References 0 publications
1
14
0
Order By: Relevance
“…Furthermore, compared to the most natural baseline of network alignment, DWSNets scale significantly better with the data. In reality, it is challenging to use this baseline due to the fact that the weight-space alignment problem is hard (Ainsworth et al, 2022). The problem is further amplified when having large input networks or large (networks) datasets.…”
Section: Analysis Of the Resultsmentioning
confidence: 99%
See 4 more Smart Citations
“…Furthermore, compared to the most natural baseline of network alignment, DWSNets scale significantly better with the data. In reality, it is challenging to use this baseline due to the fact that the weight-space alignment problem is hard (Ainsworth et al, 2022). The problem is further amplified when having large input networks or large (networks) datasets.…”
Section: Analysis Of the Resultsmentioning
confidence: 99%
“…In a fundamental work, Hecht-Nielsen (1990) observed that MLPs have permutation symmetries: swapping the order of the activations in an intermediate layer does not change the underlying function. Motivated by previous works (Hecht-Nielsen, 1990;Brea et al, 2019;Ainsworth et al, 2022) we define the weight-space of an M -layer MLP as:…”
Section: Permutation Symmetries Of Neural Networkmentioning
confidence: 99%
See 3 more Smart Citations