Second-order forward-mode optimization of recurrent neural networks for neuroscience
–Neural Information Processing Systems
A common source of anxiety for the computational neuroscience student is the question "will my recurrent neural network (RNN) model finally learn that task?". GRU or LSTM) is acceptable if it speeds up training, the RNN models trained as models of brain dynamics are subject to plausibility constraints that fundamentally exclude the usual machine learning hacks. The "vanilla" RNNs commonly used in computational neuroscience find themselves plagued by ill-conditioned loss surfaces that complicate training and significantly hinder our capacity to investigate the brain dynamics underlying complex tasks. Moreover, some tasks may require very long time horizons which backpropagation cannot handle given typical GPU memory limits. Here, we develop SOFO, a second-order optimizer that efficiently navigates loss surfaces whilst not requiring backpropagation. By relying instead on easily parallelized batched forward-mode differentiation, SOFO enjoys constant memory cost in time.
Neural Information Processing Systems
May-27-2025, 11:49:01 GMT