Generalization in Representation Models via Random Matrix Theory: Application to Recurrent Networks
Moakher, Yessin, Tiomoko, Malik, Louart, Cosme, Liao, Zhenyu
We first study the generalization error of models that use a fixed feature representation (frozen intermediate layers) followed by a trainable readout layer. This setting encompasses a range of architectures, from deep random-feature models to echo-state networks (ESNs) with recurrent dynamics. Working in the high-dimensional regime, we apply Random Matrix Theory to derive a closed-form expression for the asymptotic generalization error. We then apply this analysis to recurrent representations and obtain concise formula that characterize their performance. Surprisingly, we show that a linear ESN is equivalent to ridge regression with an exponentially time-weighted (''memory'') input covariance, revealing a clear inductive bias toward recent inputs. Experiments match predictions: ESNs win in low-sample, short-memory regimes, while ridge prevails with more data or long-range dependencies. Our methodology provides a general framework for analyzing overparameterized models and offers insights into the behavior of deep learning networks.
Nov-10-2025
- Country:
- Asia > China
- Guangdong Province > Shenzhen (0.04)
- Hong Kong (0.04)
- Hubei Province > Wuhan (0.04)
- Europe
- France
- Auvergne-Rhône-Alpes > Isère
- Grenoble (0.04)
- Île-de-France > Paris
- Paris (0.04)
- Auvergne-Rhône-Alpes > Isère
- Germany > North Rhine-Westphalia
- Cologne Region > Bonn (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- France
- Asia > China
- Genre:
- Research Report (0.64)
- Technology: