Phase Transitions between Accuracy Regimes in L2 regularized Deep Neural Networks
Ersoy, Ibrahim Talha, Wiesner, Karoline
–arXiv.org Artificial Intelligence
Increasing the L2 regularization of Deep Neural Networks (DNNs) causes a first-order phase transition into the under-parametrized phase -- the so-called onset-of learning. We explain this transition via the scalar (Ricci) curvature of the error landscape. We predict new transition points as the data complexity is increased and, in accordance with the theory of phase transitions, the existence of hysteresis effects. We confirm both predictions numerically. Our results provide a natural explanation of the recently discovered phenomenon of '\emph{grokking}' as DNN models getting stuck in a local minimum of the error surface, corresponding to a lower accuracy phase. Our work paves the way for new probing methods of the intrinsic structure of DNNs in and beyond the L2 context.
arXiv.org Artificial Intelligence
Aug-29-2025
- Country:
- Europe
- Germany > Brandenburg
- Potsdam (0.05)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Germany > Brandenburg
- North America > Canada
- Alberta > Census Division No. 15 > Improvement District No. 9 > Banff (0.04)
- Europe
- Genre:
- Research Report > New Finding (0.88)
- Technology: