“…Moreover, the last layer's weights {w k } are also aligned (i.e., equal up to a scalar factor) to the same simplex ETF, and as a result, the classification turns to be based on the nearest class center in feature space. This "neural collapse" (NC) behavior has led to many follow-up papers (Mixon et al, 2020;Lu & Steinerberger, 2022;Wojtowytsch et al, 2021;Fang et al, 2021;Zhu et al, 2021;Graf et al, 2021;Ergen & Pilanci, 2021;Zarka et al, 2021). Some of them include practical implications of the NC phenomenon, such as designing layers (multiplication by tight frames followed by soft-thresholding) that concentrate within-class features (Zarka et al, 2021) or fixing the last layer's weights to be a simplex ETF (Zhu et al, 2021).…”