XY Han, Prevalence of Neural Collapse during the terminal phase of deep learning train., 2021.10.05
The following notes are based on the seminar video by one of the original authors of the Neural Collapse paper.
Simplex equiangular tight frame (ETF): Equinorm, Equiangle, Maximal Angle.
[NC1] Variability collapse: within-class variable collapse (features collapse)
[NC2] Class mean convergence to simplex ETF: within a space where number of vectors/classes is smaller than the dimension of the space. The trad ETF is within a setting a where there are more vectors than dimensions → the definition of the frame is that the smallest eigenvalue of the vectors of the frame is greater than zero, it spans the space. Here, the spanning condition is relaxed and the structure is retained.
[NC3] Convergence to self-duality: means and classifier converge to each other.
[NC4] Behavioral convergence to nearest class center: The classifier is simply the nearest neighbor to the class mean (See theoretical proof in the paper).
What advantages do we have if we train beyond zero error?
- Improves generalization, test performance.
- Improves adversarial robustness (DeepFool metric)
Feature Engineering
Notation exposes only last-layer objects: and . Full notation
- : last-layer classifier (-dimensional vectors)
- : parameters of earlier layers Last-layer activations:
Optimizing : ‘Feature engineering’.
Neural Collapse is the preferred end-state of every successful exercise in feature engineering. Similar idea was proposed by Shannon in his 1959 paper “Probability of Error for Optimal Codes in a Gaussian Channel”.
Linear Decoding Over Gaussian Channel
Transmitter transmits codeword: .
Receiver receives: .
Linear decoder with weights and biases :
Goal: Design weights, biases and codewords such that . Here we have a norm bound on the means. So choose the best codeword so that the decoder’s job is the easiest, basically learn features in such a way that it is easier to classify as for our case.
This is similar to the feature engineering problem we have. Shannon gave bounds for what the angles between the codewords must be as the dimension of the codewords (size of the codewords) increases.
For us, we look at other things. We know that for deep nets, the signal overwhelms the noise therefore . So what is the limiting ratio that we should look at, limit exists as we go to infinity? Defined below. (this is different from what Shannon considered).
Large deviations error exponent
where and are opposing as initially noise is high and signal is low, thus misclassification occurs and as we keep training lowers essentially tending to , first time thus increasing and the rate of misclassification decreases.
Optimal Error Exponent
References
[1] Papyan, Vardan, X. Y. Han, and David L. Donoho. “Prevalence of neural collapse during the terminal phase of deep learning training.” Proceedings of the National Academy of Sciences 117.40 (2020): 24652-24663.